Creating A Galactic Plane Atlas With Amazon Web Services
|
|
|
- Blanche Dennis
- 10 years ago
- Views:
Transcription
1 Creating A Galactic Plane Atlas With Amazon Web Services G. Bruce Berriman 1*, Ewa Deelman 2, John Good 1, Gideon Juve 2, Jamie Kinney 3, Ann Merrihew 3, and Mats Rynge 2 1 Infrared Processing and Analysis Center, California Institute of Technology, Pasadena, CA 91125, USA 2 Information Sciences Institute, University of Southern California, Marina Del Rey, CA 90292, USA 3 Amazon Web Services, th Ave, Seattle, WA 98101, USA {gbb, jcg}@ipac.caltech.edu; {deelman, gideon, rynge}@isi.edu; {jkinney,amerrihew}@amazon.com Corresponding author Abstract This paper describes by example how astronomers can use cloud-computing resources offered by Amazon Web Services (AWS) to create new datasets at scale. We have created from existing surveys an atlas of the Galactic Plane at 16 wavelengths from 1 µm to 24 µm with pixels coregistered at spatial sampling of 1 arcsec. We explain how open source tools support management and operation of a virtual cluster on AWS platforms to process data at scale, and describe the technical issues that users will need to consider, such as optimization of resources, resource costs, and management of virtual machine instances. Keywords: Astronomy, Astronomy Surveys, Cloud Computing, Image Processing, Scientific Workflows, Amazon Web Services. 1. Introduction: Creating The Galactic Plane Atlas How can astronomers who are inexpert in technology take advantage of cloud computing technologies? And how can astronomers optimize performance and minimize costs under the payas-you-go tariffs for processing and storage offered by cloud computing providers? We have set out to answer these questions by using open source tools and methods to create and operate a virtual cluster on the Amazon Elastic Cloud 2 (hereafter, EC2) of Amazon Web Services (AWS) and create an imaging Atlas of the galactic plane at 16 infrared wavelengths from 1 µm to 24 µm. The Atlas has been created from input images from five major surveys, listed in Table 1 and archived at the NASA/IPAC Infrared Science Archive (IRSA), processed with the Montage image mosaic engine [1] (see also a toolkit that performs all the steps to create mosaics for a set of input files covering an area of the sky. Pixel data in all images have been transformed to 1 arcsec spatial sampling in the Cartesian projection, co-registered on the sky, and represented in Galactic coordinates. The Atlas thus appears to have been measured with a single instrument and telescope operating across the full wavelength range, incorporating all those data within Galactic coordinates l=±360 and b=±20 ; Table 1 includes the coverage of each 1
2 survey within this area. The Atlas is organized and stored in tiles 5 x 5, with tile centers separated by 4. The final output data set size has a volume of 45 TB. The organization provides a 1 overlap between tiles aimed at supporting validation. Following completion of validation, the Atlas will be archived at AWS, and made publicly accessible in spring 2014 through an Applications Programming Interface (API). Table 1: The surveys and bands included in the Galactic Plane Atlas, including output data sizes and compute times on the Amazon Cloud. Survey Bands (µm) Coverage of Atlas area (%) Output Size (TB) Compute Time 1 (1,000 s core hours) 2MASS GLIMPSE MIPSGAL MSX WISE 1.2, 1.6, , 4.5, 5.8, , 12.1, 14.6, , 4.6, 12, On the EC2 hi1.4xlarge EC2 instance. See text for details. EC2 uses virtualization technologies to offer essentially unlimited on-demand, pay-as-you-go access to compute and storage resources, which are released on completion. Astronomers are generally not skilled at managing and optimizing the environments of clusters of virtual machines, and we describe here how well open source tools can successfully manage many of these tasks on behalf of the user. Consider throughout that in the pay-as-you-go cost model, operating inefficiencies may well lead to substantially increased costs. 2. Managing Workflows For The Galactic Plane Atlas Creation of the galactic plane atlas is an example of a highly parallelizable, data driven workflow, where the output of one step becomes the input to the next, and where each step can be parallelized across as many machines as are available. Executing such a workflow on a distributed platform is 2
3 feasible but tedious [3], for it involves, among things, locating compute and storage resources, scheduling jobs and managing failures Our recommended approach is to take advantage of workflow management systems (hereafter, WMS), designed to perform these tasks. These tools take descriptions of workflows and set up and run a processing plan for an execution environment whose configuration and organization is incorporated into the workflow manager. In this way, workflow managers abstract the details of the operation of a virtual cluster away from the user. We have used in this study the Pegasus WMS [4] (see also a mature, highly-fault tolerant system used in many disciplines. A description of the workflow that is, the data flow and processing paths (including dependencies between steps) are represented in XML format as a Directed Acyclical Graph (DAG) and are created by APIs included in Pegasus. Pegasus takes this DAG and creates an executable workflow optimized for the target environment, in this case EC2. Figure 1 shows the organization of one of 16 hierarchical workflows, one for each wavelength. Pegasus is layered on top of two open source tools required to run the workflows. HTCondor ( performs job queuing and scheduling, and DAGMan (Directed Acyclic Graph Manager; ensures that jobs are submitted to HTCondor in the order implied by the dependencies encoded in the DAG. 3
4 Fig 1: A representation of the creation of one of the 16 workflows used to create the Galactic Plane Atlas. The colored circles represent one step in the processing, with the Montage components named in the legend. Each workflow contains 1,001 sub-workflows, one for each 5 x 5 tile. 3. Running the Workflows on the Amazon Web Services Cloud. Given that Montage and Pegasus are designed to run on all common Unix-based platforms, our technical approach has been to create an environment within a virtual machine in which these applications run, and subsequently replicate this machine across nodes of a cluster. (an alternative approach is to adapt applications to a MapReduce framework such as Hadoop; e.g. [2]). Creating and configuring virtual machine images and organizing them into a distributed cluster is, however, tedious and requires system administration knowledge. The tool set described above Montage, Pegasus, and HTCondor (along with dependent services) was copied and built on a virtual machine image that is loaded on to the EC2 virtual machine, which was then configured for the correct networking and system setup. Unless users have systems management experience, we strongly recommend they take advantage of existing images one for EC2 is provided on the Pegasus web page or use a service such as Puppet ( or Chef ( to construct and manage them. Creating a virtual cluster manually, by launching and configuring individual nodes is also not recommended. Tools such as Wrangler ( Automating.pdf), can easily automate the provisioning and configuration of clusters running on Amazon EC2. Figure 2 shows the architecture of the virtual cluster and storage created for the Galactic Plane Atlas. HTCondor requires the master/worker architecture shown there. The master instance is used for submitting workflows to the worker nodes for processing. The worker nodes used Amazon Simple Storage Service S3 (hereafter, S3) object storage, supported by Pegasus, to provide a centralized storage system to house input data, temporary storage for intermediate data, and secure place for the final data products. S3 is a highly efficient choice for storage as it scales with the number of clients, at the cost of some latency in each request. Ephemeral disks were used in a striped RAID-0 configuration to increase the local disk I/O performance. Processing performance was limited by the data transfer rate from IPAC. To manage data transfer rates, we used a reverse Squid caching server ( installed at ISI, as indicated in the bottom frame of Figure 3. We added and removed nodes manually as needed in response to the availability of IPAC network egress bandwidth, rather than allow the cluster to change size in response to processing demand. This provisioning strategy maintains the system at a size that the data rates can sustain. Compute nodes achieved close to 100% CPU utilization by oversubscribing the worker nodes with job. The over subscription was done in order to better 4
5 overlap the CPU intensive computations and manage the somewhat high latency in data transfers to and from S3. Thus, while Pegasus provided a robust framework for creating and executing the workflows, fully optimizing the processing required human intervention and knowledge of how to design virtual machines and how the workflow management tools. 4. Processing Costs Users should ideally perform a cost benefit analysis or benchmarking to understand which EC2 instances will meet their performance goals and budget constraints. We used the EC2 hi1.4xlarge instance type, and this consumed 318,000 core hours to complete the entire set of workflows over all 16 bands, and this would have cost $5,950 using spot pricing (essentially, bidding on unused capacity). Comparison tests show that if we had chosen to run the workload on the cc2.8xlarge instances with spot pricing, it would reduce the total cost to $2,200 based on the lower cost per core/hour and faster processor speed. These costs do not include storing the resulting data set in Amazon S3. 5
6 Fig 2: The architecture of the processing environment. In the Amazon EC2 cloud, A master instance is used to start workflows on worker nodes, which use centralized Amazon S3 storage. Outside the cloud, a squid server installed at ISI controls the rate of data transfer from IPAC to EC2. 5. Conclusions The creation of the Galactic Plane Atlas shows that Amazon EC2 is a powerful resource for astronomical computing at scale, and open source tools are invaluable for taking advantage of it. Nevertheless, managing a virtual environment, optimizing processing and controlling costs requires investment in understanding how AWS, Linux systems and the workflow tools operate. The pay-as-you go cost model requires cost benefit analyses to understand the cost and performance of AWS processing and storage options. Acknowledgements Compute resources for the project were provided by an AWS in Education Research Grant ( Storage for the Galactic Plane data set is provided by the AWS Public Data Sets program ( This paper is based upon work supported in part by the National Science Foundation under Grants (OCI , ). G. B. Berriman and J. C. Good are supported by the NASA Exoplanet Science Institute at the Infrared Processing and Analysis Center, operated by the California Institute of Technology in coordination with the Jet Propulsion Laboratory (JPL). References [1] J. C. Jacob et al. Montage: a grid portal and software toolkit for science-grade astronomical image mosaicking. Int. J. Comput. Sci. Eng. 4, 2009, [2] K. Wiley et al. Astronomical Image Processing with Hadoop. Astronomical Data Analysis Software and Systems XX. ASP Conference Proceedings, Vol. 442, 2011, Pages [3] K. R. Jackson et al. Performance and cost analysis of the Supernova factory on the Amazon AWS cloud. Scientific Programming - Science-Driven Cloud Computing, 2011, Volume 19, Pages [4] E. Deelman et al. Pegasus: a Framework for Mapping Complex Scientific Workflows onto Distributed Systems. Scientific Programming Journal, Vol 13(3), 2005, Pages
7 Bruce Berriman is a Senior Scientist at IPAC, Caltech (PhD in Astronomy, Caltech in 1983). His research centers on investigating the applicability of emerging technologies to astronomy. Ewa Deelman is a Project Leader at USC/ISI (PhD in Computer Science, RPI in 1997). Her research interests include distributed scientific environments, with emphasis on workflow management. John Good is a distributed science information systems architect at IPAC (PhD in Astronomy, U. Mass, 1983). His focus is on scalable architectures for handling extremely large datasets, especially astronomical imaging and relational catalogs and on science visualization/user interfaces. Gideon Juve is a Computer Scientist at USC/ISI (PhD in Computer Science, USC in 2012). His research focuses on enabling and optimizing scientific workflows on clusters, grids and clouds. Jamie Kinney leads the HPC and Scientific Computing team at Amazon Web Services (BS in Marine Science/Biology, University of Miami in 1996). His work focuses on the application of cloud technologies to a broad range of scientific workloads Ann Merrihew is an Amazon Web Services (AWS) Account Manager for Higher Education in California, and focused on technology and Education previously at Microsoft ( ) and Apple ( ) prior to joining AWS in early Mats Rynge is a Computer Scientist at USC/ISI (BS in Computer Science, UCLA in 2002). His research interests include distributed and high performance computing. 7
SURVEY ON THE ALGORITHMS FOR WORKFLOW PLANNING AND EXECUTION
SURVEY ON THE ALGORITHMS FOR WORKFLOW PLANNING AND EXECUTION Kirandeep Kaur Khushdeep Kaur Research Scholar Assistant Professor, Department Of Cse, Bhai Maha Singh College Of Engineering, Bhai Maha Singh
The Application of Cloud Computing to Scientific Workflows: A Study of Cost and Performance
The Application of Cloud Computing to Scientific Workflows: A Study of Cost and Performance G. Bruce Berriman, Ewa Deelman, Gideon Juve, Mats Rynge and Jens-S. Vöckler G. Bruce Berriman Infrared Processing
Data Sharing Options for Scientific Workflows on Amazon EC2
Data Sharing Options for Scientific Workflows on Amazon EC2 Gideon Juve, Ewa Deelman, Karan Vahi, Gaurang Mehta, Benjamin P. Berman, Bruce Berriman, Phil Maechling Francesco Allertsen Vrije Universiteit
Cloud Computing @ JPL Science Data Systems
Cloud Computing @ JPL Science Data Systems Emily Law, GSAW 2011 Outline Science Data Systems (SDS) Space & Earth SDSs SDS Common Architecture Components Key Components using Cloud Computing Use Case 1:
Hosted Science: Managing Computational Workflows in the Cloud. Ewa Deelman USC Information Sciences Institute
Hosted Science: Managing Computational Workflows in the Cloud Ewa Deelman USC Information Sciences Institute http://pegasus.isi.edu [email protected] The Problem Scientific data is being collected at an
Big Data on AWS. Services Overview. Bernie Nallamotu Principle Solutions Architect
on AWS Services Overview Bernie Nallamotu Principle Solutions Architect \ So what is it? When your data sets become so large that you have to start innovating around how to collect, store, organize, analyze
A Service for Data-Intensive Computations on Virtual Clusters
A Service for Data-Intensive Computations on Virtual Clusters Executing Preservation Strategies at Scale Rainer Schmidt, Christian Sadilek, and Ross King [email protected] Planets Project Permanent
Cloud computing - Architecting in the cloud
Cloud computing - Architecting in the cloud [email protected] 1 Outline Cloud computing What is? Levels of cloud computing: IaaS, PaaS, SaaS Moving to the cloud? Architecting in the cloud Best practices
Big data, big deal? Presented by: David Stanley, CTO PCI Geomatics
Big data, big deal? Presented by: David Stanley, CTO PCI Geomatics A quick word about me Ah, the early 1980s Aha! Moments in my Career First personal computer (Pet 2001, 1978) First windowing OS (SUN,
Scientific Workflow Applications on Amazon EC2
Scientific Workflow Applications on Amazon EC2 Gideon Juve, Ewa Deelman, Karan Vahi, Gaurang Mehta USC Information Sciences Institute {gideon,deelman,vahi,gmehta}@isi.edu Bruce Berriman NASA Exoplanet
Cloud Computing and Amazon Web Services
Cloud Computing and Amazon Web Services Gary A. McGilvary edinburgh data.intensive research 1 OUTLINE 1. An Overview of Cloud Computing 2. Amazon Web Services 3. Amazon EC2 Tutorial 4. Conclusions 2 CLOUD
Financial Services Grid Computing on Amazon Web Services January 2013 Ian Meyers
Financial Services Grid Computing on Amazon Web Services January 2013 Ian Meyers (Please consult http://aws.amazon.com/whitepapers for the latest version of this paper) Page 1 of 15 Contents Abstract...
Scalable Architecture on Amazon AWS Cloud
Scalable Architecture on Amazon AWS Cloud Kalpak Shah Founder & CEO, Clogeny Technologies [email protected] 1 * http://www.rightscale.com/products/cloud-computing-uses/scalable-website.php 2 Architect
The Case for Resource Sharing in Scientific Workflow Executions
The Case for Resource Sharing in Scientific Workflow Executions Ricardo Oda, Daniel Cordeiro, Rafael Ferreira da Silva 2 Ewa Deelman 2, Kelly R. Braghetto Instituto de Matemática e Estatística Universidade
Migrating a (Large) Science Database to the Cloud
The Sloan Digital Sky Survey Migrating a (Large) Science Database to the Cloud Ani Thakar Alex Szalay Center for Astrophysical Sciences and Institute for Data Intensive Engineering and Science (IDIES)
Rethinking Data Management for Big Data Scientific Workflows
Rethinking Data Management for Big Data Scientific Workflows Karan Vahi, Mats Rynge, Gideon Juve, Rajiv Mayani, Ewa Deelman Information Sciences Institute - University of Southern California Marina Del
Rethinking Data Management for Big Data Scientific Workflows
Rethinking Data Management for Big Data Scientific Workflows Karan Vahi, Mats Rynge, Gideon Juve, Rajiv Mayani, Ewa Deelman Information Sciences Institute - University of Southern California Marina Del
Resource Scalability for Efficient Parallel Processing in Cloud
Resource Scalability for Efficient Parallel Processing in Cloud ABSTRACT Govinda.K #1, Abirami.M #2, Divya Mercy Silva.J #3 #1 SCSE, VIT University #2 SITE, VIT University #3 SITE, VIT University In the
Introduction to AWS Economics
Introduction to AWS Economics Reducing Costs and Complexity May 2015 2015, Amazon Web Services, Inc. or its affiliates. All rights reserved. Notices This document is provided for informational purposes
ZADARA STORAGE. Managed, hybrid storage EXECUTIVE SUMMARY. Research Brief
ZADARA STORAGE Managed, hybrid storage Research Brief EXECUTIVE SUMMARY In 2013, Neuralytix first documented Zadara s rise to prominence in the then, fledgling integrated on-premise and in-cloud storage
Amazon EC2 Product Details Page 1 of 5
Amazon EC2 Product Details Page 1 of 5 Amazon EC2 Functionality Amazon EC2 presents a true virtual computing environment, allowing you to use web service interfaces to launch instances with a variety of
Hadoop & Spark Using Amazon EMR
Hadoop & Spark Using Amazon EMR Michael Hanisch, AWS Solutions Architecture 2015, Amazon Web Services, Inc. or its Affiliates. All rights reserved. Agenda Why did we build Amazon EMR? What is Amazon EMR?
A Brief Introduction to Apache Tez
A Brief Introduction to Apache Tez Introduction It is a fact that data is basically the new currency of the modern business world. Companies that effectively maximize the value of their data (extract value
THE DEFINITIVE GUIDE FOR AWS CLOUD EC2 FAMILIES
THE DEFINITIVE GUIDE FOR AWS CLOUD EC2 FAMILIES Introduction Amazon Web Services (AWS), which was officially launched in 2006, offers you varying cloud services that are not only cost effective, but also
Alternative Deployment Models for Cloud Computing in HPC Applications. Society of HPC Professionals November 9, 2011 Steve Hebert, Nimbix
Alternative Deployment Models for Cloud Computing in HPC Applications Society of HPC Professionals November 9, 2011 Steve Hebert, Nimbix The case for Cloud in HPC Build it in house Assemble in the cloud?
Early Cloud Experiences with the Kepler Scientific Workflow System
Available online at www.sciencedirect.com Procedia Computer Science 9 (2012 ) 1630 1634 International Conference on Computational Science, ICCS 2012 Early Cloud Experiences with the Kepler Scientific Workflow
Big data blue print for cloud architecture
Big data blue print for cloud architecture -COGNIZANT Image Area Prabhu Inbarajan Srinivasan Thiruvengadathan Muralicharan Gurumoorthy Praveen Codur 2012, Cognizant Next 30 minutes Big Data / Cloud challenges
Final Project Proposal. CSCI.6500 Distributed Computing over the Internet
Final Project Proposal CSCI.6500 Distributed Computing over the Internet Qingling Wang 660795696 1. Purpose Implement an application layer on Hybrid Grid Cloud Infrastructure to automatically or at least
Cluster, Grid, Cloud Concepts
Cluster, Grid, Cloud Concepts Kalaiselvan.K Contents Section 1: Cluster Section 2: Grid Section 3: Cloud Cluster An Overview Need for a Cluster Cluster categorizations A computer cluster is a group of
Leveraging BlobSeer to boost up the deployment and execution of Hadoop applications in Nimbus cloud environments on Grid 5000
Leveraging BlobSeer to boost up the deployment and execution of Hadoop applications in Nimbus cloud environments on Grid 5000 Alexandra Carpen-Amarie Diana Moise Bogdan Nicolae KerData Team, INRIA Outline
DynamicCloudSim: Simulating Heterogeneity in Computational Clouds
DynamicCloudSim: Simulating Heterogeneity in Computational Clouds Marc Bux, Ulf Leser {bux leser}@informatik.hu-berlin.de The 2nd international workshop on Scalable Workflow Enactment Engines and Technologies
Introduction to Cloud Computing
Introduction to Cloud Computing Cloud Computing I (intro) 15 319, spring 2010 2 nd Lecture, Jan 14 th Majd F. Sakr Lecture Motivation General overview on cloud computing What is cloud computing Services
Cloud Computing. Alex Crawford Ben Johnstone
Cloud Computing Alex Crawford Ben Johnstone Overview What is cloud computing? Amazon EC2 Performance Conclusions What is the Cloud? A large cluster of machines o Economies of scale [1] Customers use a
Network Infrastructure Services CS848 Project
Quality of Service Guarantees for Cloud Services CS848 Project presentation by Alexey Karyakin David R. Cheriton School of Computer Science University of Waterloo March 2010 Outline 1. Performance of cloud
Axceleon s CloudFuzion Turbocharges 3D Rendering On Amazon s EC2
Axceleon s CloudFuzion Turbocharges 3D Rendering On Amazon s EC2 In the movie making, visual effects and 3D animation industrues meeting project and timing deadlines is critical to success. Poor quality
COMPUTER MEASUREMENT GROUP - India Hyderabad Chapter. Strategies to Optimize Cloud Costs By Cloud Performance Monitoring
COMPUTER MEASUREMENT GROUP - India Hyderabad Chapter Strategies to Optimize Cloud Costs By Cloud Performance Monitoring October 2013 www.cmgindia.org Computer Measurement Group, India 1 About Me Credentials
Storage Solutions in the AWS Cloud. Miles Ward Enterprise Solutions Architect
Storage Solutions in the AWS Cloud Miles Ward Enterprise Solutions Architect Traditional Storage On-Premise Storage Options SAN network-attached block devices: LUNs DAS local block devices (disks) NAS
On the Performance-cost Tradeoff for Workflow Scheduling in Hybrid Clouds
On the Performance-cost Tradeoff for Workflow Scheduling in Hybrid Clouds Thiago A. L. Genez, Luiz F. Bittencourt, Edmundo R. M. Madeira Institute of Computing University of Campinas UNICAMP Av. Albert
Hadoop in the Hybrid Cloud
Presented by Hortonworks and Microsoft Introduction An increasing number of enterprises are either currently using or are planning to use cloud deployment models to expand their IT infrastructure. Big
Building your Big Data Architecture on Amazon Web Services
Building your Big Data Architecture on Amazon Web Services Abhishek Sinha @abysinha [email protected] AWS Services Deployment & Administration Application Services Compute Storage Database Networking
Cloud 101. Mike Gangl, Caltech/JPL, [email protected] 2015 California Institute of Technology. Government sponsorship acknowledged
Cloud 101 Mike Gangl, Caltech/JPL, [email protected] 2015 California Institute of Technology. Government sponsorship acknowledged Outline What is cloud computing? Cloud service models Deployment
A REVIEW PAPER ON THE HADOOP DISTRIBUTED FILE SYSTEM
A REVIEW PAPER ON THE HADOOP DISTRIBUTED FILE SYSTEM Sneha D.Borkar 1, Prof.Chaitali S.Surtakar 2 Student of B.E., Information Technology, J.D.I.E.T, [email protected] Assistant Professor, Information
Introduction to Cloud : Cloud and Cloud Storage. Lecture 2. Dr. Dalit Naor IBM Haifa Research Storage Systems. Dalit Naor, IBM Haifa Research
Introduction to Cloud : Cloud and Cloud Storage Lecture 2 Dr. Dalit Naor IBM Haifa Research Storage Systems 1 Advanced Topics in Storage Systems for Big Data - Spring 2014, Tel-Aviv University http://www.eng.tau.ac.il/semcom
Hadoop Cluster Applications
Hadoop Overview Data analytics has become a key element of the business decision process over the last decade. Classic reporting on a dataset stored in a database was sufficient until recently, but yesterday
Alfresco Enterprise on AWS: Reference Architecture
Alfresco Enterprise on AWS: Reference Architecture October 2013 (Please consult http://aws.amazon.com/whitepapers/ for the latest version of this paper) Page 1 of 13 Abstract Amazon Web Services (AWS)
Workflow Allocations and Scheduling on IaaS Platforms, from Theory to Practice
Workflow Allocations and Scheduling on IaaS Platforms, from Theory to Practice Eddy Caron 1, Frédéric Desprez 2, Adrian Mureșan 1, Frédéric Suter 3, Kate Keahey 4 1 Ecole Normale Supérieure de Lyon, France
Zadara Storage Cloud A whitepaper. @ZadaraStorage
Zadara Storage Cloud A whitepaper @ZadaraStorage Zadara delivers two solutions to its customers: On- premises storage arrays Storage as a service from 31 locations globally (and counting) Some Zadara customers
There Are Clouds In Your Future. Jeff Barr Amazon Web Services [email protected] @jeffbarr (Twitter)
There Are Clouds In Your Future Jeff Barr Amazon Web Services [email protected] @jeffbarr (Twitter) My Goals For This Talk Introduce you to cloud computing Show you what others are already doing Alert you
IBM Platform Computing Cloud Service Ready to use Platform LSF & Symphony clusters in the SoftLayer cloud
IBM Platform Computing Cloud Service Ready to use Platform LSF & Symphony clusters in the SoftLayer cloud February 25, 2014 1 Agenda v Mapping clients needs to cloud technologies v Addressing your pain
Cloud Computing using MapReduce, Hadoop, Spark
Cloud Computing using MapReduce, Hadoop, Spark Benjamin Hindman [email protected] Why this talk? At some point, you ll have enough data to run your parallel algorithms on multiple computers SPMD (e.g.,
Part V Applications. What is cloud computing? SaaS has been around for awhile. Cloud Computing: General concepts
Part V Applications Cloud Computing: General concepts Copyright K.Goseva 2010 CS 736 Software Performance Engineering Slide 1 What is cloud computing? SaaS: Software as a Service Cloud: Datacenters hardware
Cloud Computing and E-Commerce
Cloud Computing and E-Commerce Cloud Computing turns Computing Power into a Virtual Good for E-Commerrce is Implementation Partner of 4FriendsOnly.com Internet Technologies AG VirtualGoods, Koblenz, September
Challenges of Managing Scientific Workflows in High-Throughput and High- Performance Computing Environments
Challenges of Managing Scientific Workflows in High-Throughput and High- Performance Computing Environments Ewa Deelman USC Informa1on Sciences Ins1tute h8p://www.isi.edu/~deelman Funding from DOE, NSF,
Chapter 19 Cloud Computing for Multimedia Services
Chapter 19 Cloud Computing for Multimedia Services 19.1 Cloud Computing Overview 19.2 Multimedia Cloud Computing 19.3 Cloud-Assisted Media Sharing 19.4 Computation Offloading for Multimedia Services 19.5
Boas Betzler. Planet. Globally Distributed IaaS Platform Examples AWS and SoftLayer. November 9, 2015. 20014 IBM Corporation
Boas Betzler Cloud IBM Distinguished Computing Engineer for a Smarter Planet Globally Distributed IaaS Platform Examples AWS and SoftLayer November 9, 2015 20014 IBM Corporation Building Data Centers The
A Comparison of Clouds: Amazon Web Services, Windows Azure, Google Cloud Platform, VMWare and Others (Fall 2012)
1. Computation Amazon Web Services Amazon Elastic Compute Cloud (Amazon EC2) provides basic computation service in AWS. It presents a virtual computing environment and enables resizable compute capacity.
Cloudera Enterprise Reference Architecture for Google Cloud Platform Deployments
Cloudera Enterprise Reference Architecture for Google Cloud Platform Deployments Important Notice 2010-2015 Cloudera, Inc. All rights reserved. Cloudera, the Cloudera logo, Cloudera Impala, Impala, and
Putchong Uthayopas, Kasetsart University
Putchong Uthayopas, Kasetsart University Introduction Cloud Computing Explained Cloud Application and Services Moving to the Cloud Trends and Technology Legend: Cluster computing, Grid computing, Cloud
Service Organization Controls 3 Report
Service Organization Controls 3 Report Report on the Amazon Web Services System Relevant to Security For the Period April 1, 2013 March 31, 2014 Ernst & Young LLP Suite 1600 560 Mission Street San Francisco,
Archiving and Sharing Big Data Digital Repositories, Libraries, Cloud Storage
Archiving and Sharing Big Data Digital Repositories, Libraries, Cloud Storage Cyrus Shahabi, Ph.D. Professor of Computer Science & Electrical Engineering Director, Integrated Media Systems Center (IMSC)
Data Semantics Aware Cloud for High Performance Analytics
Data Semantics Aware Cloud for High Performance Analytics Microsoft Future Cloud Workshop 2011 June 2nd 2011, Prof. Jun Wang, Computer Architecture and Storage System Laboratory (CASS) Acknowledgement
Cloud Computing. Chapter 1 Introducing Cloud Computing
Cloud Computing Chapter 1 Introducing Cloud Computing Learning Objectives Understand the abstract nature of cloud computing. Describe evolutionary factors of computing that led to the cloud. Describe virtualization
AN IMPLEMENTATION OF E- LEARNING SYSTEM IN PRIVATE CLOUD
AN IMPLEMENTATION OF E- LEARNING SYSTEM IN PRIVATE CLOUD M. Lawanya Shri 1, Dr. S. Subha 2 1 Assistant Professor,School of Information Technology and Engineering, Vellore Institute of Technology, Vellore-632014
Cloud Computing: Computing as a Service. Prof. Daivashala Deshmukh Maharashtra Institute of Technology, Aurangabad
Cloud Computing: Computing as a Service Prof. Daivashala Deshmukh Maharashtra Institute of Technology, Aurangabad Abstract: Computing as a utility. is a dream that dates from the beginning from the computer
Deploying Splunk on Amazon Web Services
Copyright 2014 Splunk Inc. Deploying Splunk on Amazon Web Services Simeon Yep Senior Manager, Business Development Technical Services Roy Arsan Senior SoHware Engineer Disclaimer During the course of this
Developing Scalable Smart Grid Infrastructure to Enable Secure Transmission System Control
Developing Scalable Smart Grid Infrastructure to Enable Secure Transmission System Control EP/K006487/1 UK PI: Prof Gareth Taylor (BU) China PI: Prof Yong-Hua Song (THU) Consortium UK Members: Brunel University
Financial Services Grid Computing on Amazon Web Services. January, 2016
Financial Services Grid Computing on Amazon Web Services January, 2016 2015, Amazon Web Services, Inc. or its affiliates. All rights reserved. Notices This document is provided for informational purposes
Cloud computing doesn t yet have a
The Case for Cloud Computing Robert L. Grossman University of Illinois at Chicago and Open Data Group To understand clouds and cloud computing, we must first understand the two different types of clouds.
Cloud Computing Deja Vu
Cloud Computing Deja Vu Hengming Zou, Ph.D. Shanghai Jiao Tong University Time Is Late An auto company need to know the price of front bumper And want it right now So it reconfigures some server to run
Amazon Web Services Primer. William Strickland COP 6938 Fall 2012 University of Central Florida
Amazon Web Services Primer William Strickland COP 6938 Fall 2012 University of Central Florida AWS Overview Amazon Web Services (AWS) is a collection of varying remote computing provided by Amazon.com.
19.10.11. Amazon Elastic Beanstalk
19.10.11 Amazon Elastic Beanstalk A Short History of AWS Amazon started as an ECommerce startup Original architecture was restructured to be more scalable and easier to maintain Competitive pressure for
