Early Cloud Experiences with the Kepler Scientific Workflow System
|
|
|
- Samantha McKinney
- 9 years ago
- Views:
Transcription
1 Available online at Procedia Computer Science 9 (2012 ) International Conference on Computational Science, ICCS 2012 Early Cloud Experiences with the Kepler Scientific Workflow System Jianwu Wang, Ilkay Altintas San Diego Supercomputer Center, UCSD, 9500 Gilman Drive, MC 0505, La Jolla, CA 92093, U.S.A. Abstract With the increasing popularity of the Cloud computing, there are more and more requirements for scientific workflows to utilize Cloud resources. In this paper, we present our preliminary work and experiences on enabling the interaction between the Kepler scientific workflow system and the Amazon Elastic Compute Cloud (EC2). A set of EC2 actors and Kepler Amazon Machine Images are introduced with the discussion on their different usage modes. Through two bioinformatics usecases, we demonstrate the capability of our work for both Cloud resource coordination and workflow execution on virtual Cloud resources. Keywords: Scientific workflows, data-intensive, bioinformatics, Amazon EC2, Cloud computing 1. Introduction Because of its virtualization, abundance and scalability characteristics, Cloud is becoming a popular environment for both scientific and business applications. For example, Amazon Elastic Compute Cloud (Amazon EC2) 1 provides infrastructure as a service (IaaS), which includes virtualized hardware (including storage and network) and pre-configured software stack. Users can start one or more virtual instances from one Amazon Machine Image (AMI), and administrate them as their own machines. Users can also get storage service from Amazon Simple Storage Service (S3) 2 or Elastic Block Store (EBS) 3. Instead of making investments to purchase their own computing resources, users can start using the available Cloud resources for compute and storage instantly via these services and pay as they go based on their usage. The popularity of the Cloud computing introduced new requirements for scientific workflows [1]. First, workflow users should be able to utilize available Cloud resources and packages since more and more domain-specific toolkits are available on the Cloud, especially on Amazon EC2. As a typical instance, Bio-Linux project 4 provides an AMI containing more than 500 bioinformatics programs. By initiating the Bio-Linux AMI on EC2, users can run these Corresponding author. Tel.: ; fax: address: {jianwu, altintas}@sdsc.edu (Jianwu Wang, Ilkay Altintas) 1 Amazon EC2: Amazon S3: Amazon EBS: Bio-Linux Project: Published by Elsevier Ltd. doi: /j.procs
2 Jianwu Wang and Ilkay Altintas / Procedia Computer Science 9 ( 2012 ) programs directly without a separate installation. A complex application may have to execute toolkits on multiple AMIs. Workflow systems could provide a way for scientists to utilize these available tools by scheduling tasks with proper dependency control. The tasks include Cloud instance management (e.g., initiation and termination), data transfer between Cloud instances, and program execution on Cloud instances. Another new requirement is dataintensive workflow application execution on the Cloud. For data-intensive applications, localities of data and programs are important for the overall performance. With virtualization and other techniques, workflow applications could get both data locality and program locality. We will discuss this in more detail in Section 3. In this paper, we present our early work utilizing Amazon EC2 for the Kepler scientific workflow System 5 [2, 3]. The rest of this paper is organized as follows. In Section 2, we describe our Kepler EC2 Actors and AMIs to to interact with Amazon Cloud. Different usage modes of these Actors and AMIs are discussed in Section 3. Two workflow usecases in bioinformatics domain are demostrated in Section 4. In Section 5, we discuss our conclusions and plans for future work. 2. Kepler Amazon EC2 Actors and Amazon Machine Images Kepler EC2 Actors. We have implemented an EC2 module in Kepler which contains a set of EC2 actors. These actors can be used to manage EC2 virtual instances in Amazon Cloud environment and attach EBS Volumes. Users can use these actors to start, stop, run or terminate EC2 instances. Through these EC2 actors and other actors in Kepler, users can easily build workflows to run their applications on EC2 Cloud resources. One EC2 workflow can link a variety of applications residing on different platforms by connecting multiple AMIs, e.g., Hadoop 6 and Linux AMIs. Furthermore, through the Kepler user interface, researchers could analyze bottlenecks of their processes and accelerate their execution by switching to better AMI instance types or by activating multiple parallel AMI instances for their tasks. Figure 1: Configurations of the Kepler RunEC2Instance actor. Figure 1 shows the configuration parameters for the RunEC2Instance actor. To run an EC2 instance, users need to provide account information including Access Key, Secret Key and credential Key Pair file. Image ID parameter indicates the AMI ID for the instance. Users can also configure Instance Type, Availability Zone, and Minimal Instance Number for the instance they want to run. Using these configuration parameter settings, the RunEC2Instance actor interacts with Amazon EC2 Web service in order to get configured instances. If the actor can successfully get an instance, the information of the instance, e.g., URL and ID, will be sent out through its output. Otherwise, the actor will pop up error message dialogue explaining why it cannot run successfully. Kepler Amazon Machine Images and EBS Volumes. We created a few demonstrative Amazon Machine Images and EBS Volumes for Kepler. The Images and Volumes contain the Kepler system, demonstrative Kepler workflows and third-party bioinformatics tools including the BLAST tool [4]. The difference between Kepler Images and Volumes is that Kepler Images can be used directly to run virtual instances, while Kepler Volumes have to be attached 5 Kepler website: Hadoop Project:
3 1632 Jianwu Wang and Ilkay Altintas / Procedia Computer Science 9 ( 2012 ) to running instances. These Kepler Images and Volumes can be used to provide scalable and virtualized workflow execution engine. For example, we discussed how to build a virtual Hadoop and Oracle Grid Engine 7 cluster with an attached Kepler Volume to accelerate certain Kepler workflow executions in [5]. Virtual cluster could also be built directly based on a Kepler Image using toolkits like StarCluster Usage Modes The Kepler EC2 actors and Images introduced in Section 2 could be used in different usage modes. We explain the requirements and capabilities of each usage mode based on example scenarios below. 1) Kepler EC2 Actors + Third Party AMIs : If users want to access tools on third party AMIs, such as the Bio-Linux Image, they can build workflows to access the related AMIs using the EC2 actors. Some other actors, e.g., SSHExecution and SSHFileCopier, are also needed to transfer data and call third party programs remotely. Dependencies can be easily depicted in the workflow. An example for this usage mode is discussed as the first usecase in Section 4. 2) Kepler EC2 Actors + Kepler AMI: If users want to run their existing local workflows on EC2, they can build an auxiliary workflow with EC2 actors to call their existing workflows on the virtual Kepler EC2 instance. For example, a user might have a workflow working locally, and want to run it on a faster compute node. In this usage mode, the user can find a proper instance type provided by Amazon for the Kepler AMI and configure it in the EC2 actors of the auxiliary workflow. The auxiliary workflow will initiate a Kepler virtual instance for this instance type from the Kepler AMI, upload his/her original workflows to the instance. Then the auxiliary workflow can remotely call the Kepler execution engine to execute uploaded workflows on the instance. After the execution on the instance, the results can be copied back to local machine. 3) Kepler EC2 Actors + Kepler AMI + Third Party AMIs: If users already have workflows for the first usage mode, they might want to run them via a virtualized Kepler execution engine. An auxiliary workflow can be built to initiate a Kepler virtual instance from the Kepler AMI, upload the existing workflows and run them on the instance. Then the uploaded workflows will use the EC2 actors in the workflows to launch other instances and run tools there. This usage mode can be seen as a special combination of the above two. Particularly in this mode, user account information needs to be uploaded to the Kepler virtual instance to launch other virtual instances. An example for this usage mode is discussed as the second usecase in Section 4. 4) Virtual Cluster based on Kepler AMI: This mode is suitable for distributed parallel workflow execution on the Cloud. Amazon EC2 provisions Cluster Compute Instance type and there are toolkits like StarCluster to help setting up virtual clusters based on the Kepler Image. Using such toolkits, a virtual cluster can be generated with Kepler installed on each node and fast inter-connection. This usage mode fits well with the framework for distributed data-parallel workflow execution in Kepler [6]. By utilizing distributed data-parallel execution engines like Hadoop, data can be partitioned and distributed on the virtual cluster. So with both data and Kepler execution engine located locally on each compute node, the program and data transfer among nodes could be minimized and better performance could be achieved for data-intensive workflow applications. Besides the above typically usage modes, there could be other usage modes based on particular requirements. For instance, a user could launch Kepler virtual instances using the EC2 toolkit and directly login to the instances to run workflows. 4. Bioinformatics Usecases To validate the proposed work in scientific applications, two experimental bioinformatics workflows running the BLAST tool [4] are shown in Figures 2 and 3. The BLAST tool detects the similarities between biological sequence datasets, and is one of the most widely used tools in bioinformatics. 7 Oracle Grid Engine System: StarCluster Toolkit:
4 Jianwu Wang and Ilkay Altintas / Procedia Computer Science 9 ( 2012 ) Figure 2: A Kepler workflow running BLAST on a third-party virtual instance. Figure 2 shows a Kepler workflow that runs BLAST on EC2 via the first usage mode in Section 3. The workflow first runs a virtual instance from the Bio-Linux Image using the RunEC2Instance actor. Because EC2 virtual instance might take a while to be ready for user access, the workflow uses a loop to monitor its accessibility. After the virtual instance is accessible, two SSHFileCopier actors are employed to copy the reference database and query sequence file to the Bio-Linux instance. Then BLAST tool is executed on the instance via a SSHExecute actor. After the execution is done, its output is copied back to the local machine via another SSHFileCopier actor. Figure 3: A Kepler workflow running BLAST on a Kepler virtual instance and a third-party virtual instance. Figure 3 shows a Kepler workflow that runs BLAST on EC2 via the third usage mode in Section 3. First, this workflow starts a virtual instance of a Kepler AMI. Then similar to the above workflow, it needs to wait until the virtual instance is accessible. After the virtual instance is running, user credential file to use EC2 resources and the workflow shown in Figure 2 are uploaded to the Kepler EC2 instance. The credential file copy is necessary to start another virtual EC2 instance. Then the Kepler engine on the Kepler instance is run in batch mode to execute the
5 1634 Jianwu Wang and Ilkay Altintas / Procedia Computer Science 9 ( 2012 ) uploaded workflow on another EC2 instance based on the Bio-Linux Image. After the execution, the uploaded user credential file is deleted for account security. In the end, the output file is downloaded to the local machine. We note that domain scientists do not have to build workflows from scratch like the workflows discussed above for all their requirements. We can build generic workflow templates or compose the above workflows into composite actors, only leaving parameters, such as program name and data path, for scientists to configure for their different requirements. 5. Conclusions and Future Work Utilizing advantages of Cloud computing, scientific workflows could serve scientific application easier or more efficiently. In this paper, we discuss our early Cloud experiments with the Kepler scientific workflow system, which includes our EC2 actors, Kepler AMI and their usage discussion. Through bioinformatics usecases, we demonstrate how to use our Kepler EC2 actors and Images to coordinate Cloud resources in a workflow and how to execute workflows on virtual Cloud resources. The Kepler EC2 actors are in a separate module in the Kepler repository. We will release this module in a new Kepler suite. Along with the release, the Kepler Images and EBS Volumes will also be updated. We will publish the information of the Images and Volumes in the Kepler project website. With the promising results so far, we will improve our work in the future in several directions. First, we will compare the different usage modes in Section 3 through experiments. We also plan to study how to optimize dataintensive workflow execution on EC2. 6. Acknowledgments The authors would like to thank the rest of Kepler and CAMERA teams for their collaboration. This work was supported by NSF SDCI Award OCI for Kepler/CORE, NSF ABI Award DBI for biokepler, the Gordon and Betty Moore Foundation award to Calit2 at UCSD for CAMERA and an SDSC Triton Research Opportunities grant. We also thank the AWS in Education Research Grants from Amazon.com, Inc for EC2 usage credit. References [1] G. Juve, E. Deelman, Scientific workflows and clouds, Crossroads 16 (2010) doi: [2] I. Altintas, C. Berkley, E. Jaeger, M. Jones, B. Ludaescher, S. Mock, Kepler: An extensible system for design and execution of scientific workflows, in: Proceedings of 16th International Conference on Scientific and Statistical Database Management (SSDBM 2004), Santorini Island, Greece, [3] B. Ludaescher, I. Altintas, C. Berkley, D. Higgins, E. Jaeger-Frank, M. Jones, E. A. Lee, J. Tao, Y. Zhao, Scientific workflow management and the Kepler system, Concurrency and Computation: Practice & Experience, Special Issue on Scientific Workflows 18 (10) (2006) [4] S. F. Altschul, W. Gish, W. Miller, E. W. Myers, D. J. Lipman, Basic Local Alignment Search Tool, Journal of Molecular Biology 215 (3) (1990) doi: /s (05) [5] J. Wang, P. Korambath, I. Altintas, A physical and virtual compute cluster resource load balancing approach to data-parallel scientific workflow scheduling, in: Proceedings of 2011 IEEE World Congress on Services (SERVICES 2011), 2011, pp doi: /services [6] J. Wang, D. Crawl, I. Altintas, A framework for distributed data-parallel execution in the Kepler scientific workflow system, in: Proceedings of 1st International Workshop on Advances in the Kepler Scientific Workflow System and Its Applications, ICCS 2012, 2012.
A Framework for Distributed Data-Parallel Execution in the Kepler Scientific Workflow System
Available online at www.sciencedirect.com Procedia Computer Science 9 (2012 ) 1620 1629 International Conference on Computational Science, ICCS 2012 A Framework for Distributed Data-Parallel Execution
SURVEY ON SCIENTIFIC DATA MANAGEMENT USING HADOOP MAPREDUCE IN THE KEPLER SCIENTIFIC WORKFLOW SYSTEM
SURVEY ON SCIENTIFIC DATA MANAGEMENT USING HADOOP MAPREDUCE IN THE KEPLER SCIENTIFIC WORKFLOW SYSTEM 1 KONG XIANGSHENG 1 Department of Computer & Information, Xinxiang University, Xinxiang, China E-mail:
Comparison of Distributed Data-Parallelization Patterns for Big Data Analysis: A Bioinformatics Case Study
Comparison of Distributed Data-Parallelization Patterns for Big Data Analysis: A Bioinformatics Case Study Jianwu Wang, Daniel Crawl, Ilkay Altintas San Diego Supercomputer Center University of California,
How To Build A Workflow As A Service In The Cloud (Wfaas)
Procedia Computer Science Volume 29, 2014, Pages 546 556 ICCS 2014. 14th International Conference on Computational Science Workflow as a Service in the Cloud: Architecture and Scheduling Algorithms Jianwu
Using SUSE Studio to Build and Deploy Applications on Amazon EC2. Guide. Solution Guide Cloud Computing. www.suse.com
Using SUSE Studio to Build and Deploy Applications on Amazon EC2 Guide Solution Guide Cloud Computing Cloud Computing Solution Guide Using SUSE Studio to Build and Deploy Applications on Amazon EC2 Quickly
SURVEY ON THE ALGORITHMS FOR WORKFLOW PLANNING AND EXECUTION
SURVEY ON THE ALGORITHMS FOR WORKFLOW PLANNING AND EXECUTION Kirandeep Kaur Khushdeep Kaur Research Scholar Assistant Professor, Department Of Cse, Bhai Maha Singh College Of Engineering, Bhai Maha Singh
Deploying Kepler Workflows as Services on a Cloud Infrastructure for Smart Manufacturing
Procedia Computer Science Volume 29, 2014, Pages 2254 2259 ICCS 2014. 14th International Conference on Computational Science Deploying Kepler Workflows as Services on a Cloud Infrastructure for Smart Manufacturing
Data Semantics Aware Cloud for High Performance Analytics
Data Semantics Aware Cloud for High Performance Analytics Microsoft Future Cloud Workshop 2011 June 2nd 2011, Prof. Jun Wang, Computer Architecture and Storage System Laboratory (CASS) Acknowledgement
Online Backup Guide for the Amazon Cloud: How to Setup your Online Backup Service using Vembu StoreGrid Backup Virtual Appliance on the Amazon Cloud
Online Backup Guide for the Amazon Cloud: How to Setup your Online Backup Service using Vembu StoreGrid Backup Virtual Appliance on the Amazon Cloud Here is a step-by-step set of instructions to get your
ST 810, Advanced computing
ST 810, Advanced computing Eric B. Laber & Hua Zhou Department of Statistics, North Carolina State University January 30, 2013 Supercomputers are expensive. Eric B. Laber, 2011, while browsing the internet.
19.10.11. Amazon Elastic Beanstalk
19.10.11 Amazon Elastic Beanstalk A Short History of AWS Amazon started as an ECommerce startup Original architecture was restructured to be more scalable and easier to maintain Competitive pressure for
HDFS Cluster Installation Automation for TupleWare
HDFS Cluster Installation Automation for TupleWare Xinyi Lu Department of Computer Science Brown University Providence, RI 02912 [email protected] March 26, 2014 Abstract TupleWare[1] is a C++ Framework
Cloud Computing and Amazon Web Services
Cloud Computing and Amazon Web Services Gary A. McGilvary edinburgh data.intensive research 1 OUTLINE 1. An Overview of Cloud Computing 2. Amazon Web Services 3. Amazon EC2 Tutorial 4. Conclusions 2 CLOUD
Chapter 9 PUBLIC CLOUD LABORATORY. Sucha Smanchat, PhD. Faculty of Information Technology. King Mongkut s University of Technology North Bangkok
CLOUD COMPUTING PRACTICE 82 Chapter 9 PUBLIC CLOUD LABORATORY Hand on laboratory based on AWS Sucha Smanchat, PhD Faculty of Information Technology King Mongkut s University of Technology North Bangkok
Cloud Computing Solutions for Genomics Across Geographic, Institutional and Economic Barriers
Cloud Computing Solutions for Genomics Across Geographic, Institutional and Economic Barriers Ntinos Krampis Asst. Professor J. Craig Venter Institute [email protected] http://www.jcvi.org/cms/about/bios/kkrampis/
Zend Server Amazon AMI Quick Start Guide
Zend Server Amazon AMI Quick Start Guide By Zend Technologies www.zend.com Disclaimer This is the Quick Start Guide for The Zend Server Zend Server Amazon Machine Image The information in this document
Scientific and Technical Applications as a Service in the Cloud
Scientific and Technical Applications as a Service in the Cloud University of Bern, 28.11.2011 adapted version Wibke Sudholt CloudBroker GmbH Technoparkstrasse 1, CH-8005 Zurich, Switzerland Phone: +41
Outline. High Performance Computing (HPC) Big Data meets HPC. Case Studies: Some facts about Big Data Technologies HPC and Big Data converging
Outline High Performance Computing (HPC) Towards exascale computing: a brief history Challenges in the exascale era Big Data meets HPC Some facts about Big Data Technologies HPC and Big Data converging
Background on Elastic Compute Cloud (EC2) AMI s to choose from including servers hosted on different Linux distros
David Moses January 2014 Paper on Cloud Computing I Background on Tools and Technologies in Amazon Web Services (AWS) In this paper I will highlight the technologies from the AWS cloud which enable you
VMUnify EC2 Gateway Guide
VMUnify EC2 Gateway Guide Version 2.8.1 Copyright Information This document is the exclusive property of Mindtree limited (Mindtree); the recipient agrees that they may not copy, transmit, use or disclose
Amazon EC2 Product Details Page 1 of 5
Amazon EC2 Product Details Page 1 of 5 Amazon EC2 Functionality Amazon EC2 presents a true virtual computing environment, allowing you to use web service interfaces to launch instances with a variety of
Manjrasoft Market Oriented Cloud Computing Platform
Manjrasoft Market Oriented Cloud Computing Platform Innovative Solutions for 3D Rendering Aneka is a market oriented Cloud development and management platform with rapid application development and workload
SURFsara HPC Cloud Workshop
SURFsara HPC Cloud Workshop doc.hpccloud.surfsara.nl UvA workshop 2016-01-25 UvA HPC Course Jan 2016 Anatoli Danezi, Markus van Dijk [email protected] Agenda Introduction and Overview (current
Available online at www.sciencedirect.com Available online at www.sciencedirect.com
Available online at www.sciencedirect.com Available online at www.sciencedirect.com Physics Physics Procedia Procedia 00 (2011) 24 (2012) 000 000 2293 2297 Physics Procedia www.elsevier.com/locate/procedia
Alfresco Enterprise on AWS: Reference Architecture
Alfresco Enterprise on AWS: Reference Architecture October 2013 (Please consult http://aws.amazon.com/whitepapers/ for the latest version of this paper) Page 1 of 13 Abstract Amazon Web Services (AWS)
2) Xen Hypervisor 3) UEC
5. Implementation Implementation of the trust model requires first preparing a test bed. It is a cloud computing environment that is required as the first step towards the implementation. Various tools
Sistemi Operativi e Reti. Cloud Computing
1 Sistemi Operativi e Reti Cloud Computing Facoltà di Scienze Matematiche Fisiche e Naturali Corso di Laurea Magistrale in Informatica Osvaldo Gervasi [email protected] 2 Introduction Technologies
Cloud BioLinux: Pre-configured and On-demand Bioinformatics Computing for the Genomics Community
Cloud BioLinux: Pre-configured and On-demand Bioinformatics Computing for the Genomics Community Ntinos Krampis Asst. Professor J. Craig Venter Institute [email protected] http://www.jcvi.org/cms/about/bios/kkrampis/
Chapter 7. Using Hadoop Cluster and MapReduce
Chapter 7 Using Hadoop Cluster and MapReduce Modeling and Prototyping of RMS for QoS Oriented Grid Page 152 7. Using Hadoop Cluster and MapReduce for Big Data Problems The size of the databases used in
Archiving, Indexing and Accessing Web Materials: Solutions for large amounts of data
Archiving, Indexing and Accessing Web Materials: Solutions for large amounts of data David Minor 1, Reagan Moore 2, Bing Zhu, Charles Cowart 4 1. (88)4-104 [email protected] San Diego Supercomputer Center
CONDOR CLUSTERS ON EC2
CONDOR CLUSTERS ON EC2 Val Hendrix, Roberto A. Vitillo Lawrence Berkeley National Lab ATLAS Cloud Computing R & D 1 INTRODUCTION This is our initial work on investigating tools for managing clusters and
Migration Scenario: Migrating Batch Processes to the AWS Cloud
Migration Scenario: Migrating Batch Processes to the AWS Cloud Produce Ingest Process Store Manage Distribute Asset Creation Data Ingestor Metadata Ingestor (Manual) Transcoder Encoder Asset Store Catalog
Building Platform as a Service for Scientific Applications
Building Platform as a Service for Scientific Applications Moustafa AbdelBaky [email protected] Rutgers Discovery Informa=cs Ins=tute (RDI 2 ) The NSF Cloud and Autonomic Compu=ng Center Department
Introduction to Hadoop
Introduction to Hadoop 1 What is Hadoop? the big data revolution extracting value from data cloud computing 2 Understanding MapReduce the word count problem more examples MCS 572 Lecture 24 Introduction
Approaches for Cloud and Mobile Computing
Joint CLEEN and ACROSS Workshop on Cloud Technology and Energy Efficiency in Mobile Communications at EUCNC 15, Paris, France - 29 June, 2015 Interoperable Data Management Approaches for Cloud and Mobile
Science Clouds: Early Experiences in Cloud Computing for Scientific Applications Kate Keahey and Tim Freeman
Science Clouds: Early Experiences in Cloud Computing for Scientific Applications Kate Keahey and Tim Freeman About this document The Science Clouds provide EC2-style cycles to scientific projects. This
Sriram Krishnan, Ph.D. [email protected]
Sriram Krishnan, Ph.D. [email protected] (Re-)Introduction to cloud computing Introduction to the MapReduce and Hadoop Distributed File System Programming model Examples of MapReduce Where/how to run MapReduce
Cloud Computing. Adam Barker
Cloud Computing Adam Barker 1 Overview Introduction to Cloud computing Enabling technologies Different types of cloud: IaaS, PaaS and SaaS Cloud terminology Interacting with a cloud: management consoles
ArcGIS for Server: In the Cloud
DevSummit DC February 11, 2015 Washington, DC ArcGIS for Server: In the Cloud Bonnie Stayer, Esri Session Outline Cloud Overview - Benefits - Types of clouds ArcGIS in AWS - Cloud Builder - Maintenance
Building Success on Acquia Cloud:
Building Success on Acquia Cloud: 10 Layers of PaaS TECHNICAL Guide Table of Contents Executive Summary.... 3 Introducing the 10 Layers of PaaS... 4 The Foundation: Five Layers of PaaS Infrastructure...
Increased Security, Greater Agility, Lower Costs for AWS DELPHIX FOR AMAZON WEB SERVICES WHITE PAPER
Increased Security, Greater Agility, Lower Costs for AWS DELPHIX FOR AMAZON WEB SERVICES TABLE OF CONTENTS Introduction... 3 Overview: Delphix Virtual Data Platform... 4 Delphix for AWS... 5 Decrease the
Cloud computing - Architecting in the cloud
Cloud computing - Architecting in the cloud [email protected] 1 Outline Cloud computing What is? Levels of cloud computing: IaaS, PaaS, SaaS Moving to the cloud? Architecting in the cloud Best practices
Cloud Computing: Computing as a Service. Prof. Daivashala Deshmukh Maharashtra Institute of Technology, Aurangabad
Cloud Computing: Computing as a Service Prof. Daivashala Deshmukh Maharashtra Institute of Technology, Aurangabad Abstract: Computing as a utility. is a dream that dates from the beginning from the computer
IaaS Configuration for Cloud Platforms
vrealize Automation 6.2.3 This document supports the version of each product listed and supports all subsequent versions until the document is replaced by a new edition. To check for more recent editions
Aneka Dynamic Provisioning
MANJRASOFT PTY LTD Aneka Aneka 2.0 Manjrasoft 10/22/2010 This document describes the dynamic provisioning features implemented in Aneka and how it is possible to leverage dynamic resources for scaling
Cloud-Based Big Data Analytics in Bioinformatics
Cloud-Based Big Data Analytics in Bioinformatics Presented By Cephas Mawere Harare Institute of Technology, Zimbabwe 1 Introduction 2 Big Data Analytics Big Data are a collection of data sets so large
Features of AnyShare
of AnyShare of AnyShare CONTENT Brief Introduction of AnyShare... 3 Chapter 1 Centralized Management... 5 1.1 Operation Management... 5 1.2 User Management... 5 1.3 User Authentication... 6 1.4 Roles...
AWS Service Catalog. User Guide
AWS Service Catalog User Guide AWS Service Catalog: User Guide Copyright 2016 Amazon Web Services, Inc. and/or its affiliates. All rights reserved. Amazon's trademarks and trade dress may not be used in
AWS Account Setup and Services Overview
AWS Account Setup and Services Overview 1. Purpose of the Lab Understand definitions of various Amazon Web Services (AWS) and their use in cloud computing based web applications that are accessible over
Cloud BioLinux: Pre-configured and On-demand Bioinformatics Computing for the Genomics Community
Cloud BioLinux: Pre-configured and On-demand Bioinformatics Computing for the Genomics Community Ntinos Krampis Asst. Professor J. Craig Venter Institute [email protected] http://www.jcvi.org/cms/about/bios/kkrampis/
A programming model in Cloud: MapReduce
A programming model in Cloud: MapReduce Programming model and implementation developed by Google for processing large data sets Users specify a map function to generate a set of intermediate key/value
Developing Scalable Smart Grid Infrastructure to Enable Secure Transmission System Control
Developing Scalable Smart Grid Infrastructure to Enable Secure Transmission System Control EP/K006487/1 UK PI: Prof Gareth Taylor (BU) China PI: Prof Yong-Hua Song (THU) Consortium UK Members: Brunel University
Amazon Web Services Primer. William Strickland COP 6938 Fall 2012 University of Central Florida
Amazon Web Services Primer William Strickland COP 6938 Fall 2012 University of Central Florida AWS Overview Amazon Web Services (AWS) is a collection of varying remote computing provided by Amazon.com.
HYBRID WORKFLOW POLICY MANAGEMENT FOR HEART DISEASE IDENTIFICATION DONG-HYUN KIM *1, WOO-RAM JUNG 1, CHAN-HYUN YOUN 1
HYBRID WORKFLOW POLICY MANAGEMENT FOR HEART DISEASE IDENTIFICATION DONG-HYUN KIM *1, WOO-RAM JUNG 1, CHAN-HYUN YOUN 1 1 Department of Information and Communications Engineering, Korea Advanced Institute
Cloud-pilot.doc 12-12-2010 SA1 Marcus Hardt, Marcin Plociennik, Ahmad Hammad, Bartek Palak E U F O R I A
Identifier: Date: Activity: Authors: Status: Link: Cloud-pilot.doc 12-12-2010 SA1 Marcus Hardt, Marcin Plociennik, Ahmad Hammad, Bartek Palak E U F O R I A J O I N T A C T I O N ( S A 1, J R A 3 ) F I
Automated Application Provisioning for Cloud
Automated Application Provisioning for Cloud Application Provisioning in Cloud requires mechanism to automate and repeat as and when it requires. This is mainly because the building blocks of an IT infrastructure
PACE Predictive Analytics Center of Excellence @ San Diego Supercomputer Center, UCSD. Natasha Balac, Ph.D.
PACE Predictive Analytics Center of Excellence @ San Diego Supercomputer Center, UCSD Natasha Balac, Ph.D. Brief History of SDSC 1985-1997: NSF national supercomputer center; managed by General Atomics
Implementing Graph Pattern Mining for Big Data in the Cloud
Implementing Graph Pattern Mining for Big Data in the Cloud Chandana Ojah M.Tech in Computer Science & Engineering Department of Computer Science & Engineering, PES College of Engineering, Mandya [email protected]
SURFsara HPC Cloud Workshop
SURFsara HPC Cloud Workshop www.cloud.sara.nl Tutorial 2014-06-11 UvA HPC and Big Data Course June 2014 Anatoli Danezi, Markus van Dijk [email protected] Agenda Introduction and Overview (current
AN IMPLEMENTATION OF E- LEARNING SYSTEM IN PRIVATE CLOUD
AN IMPLEMENTATION OF E- LEARNING SYSTEM IN PRIVATE CLOUD M. Lawanya Shri 1, Dr. S. Subha 2 1 Assistant Professor,School of Information Technology and Engineering, Vellore Institute of Technology, Vellore-632014
A Scalable Network Monitoring and Bandwidth Throttling System for Cloud Computing
A Scalable Network Monitoring and Bandwidth Throttling System for Cloud Computing N.F. Huysamen and A.E. Krzesinski Department of Mathematical Sciences University of Stellenbosch 7600 Stellenbosch, South
Best Practices for Using MySQL in the Cloud
Best Practices for Using MySQL in the Cloud Luis Soares, Sr. Software Engineer, MySQL Replication, Oracle Lars Thalmann, Director Replication, Backup, Utilities and Connectors THE FOLLOWING IS INTENDED
Getting Started with Amazon EC2 Management in Eclipse
Getting Started with Amazon EC2 Management in Eclipse Table of Contents Introduction... 4 Installation... 4 Prerequisites... 4 Installing the AWS Toolkit for Eclipse... 4 Retrieving your AWS Credentials...
Best Practices for Sharing Imagery using Amazon Web Services. Peter Becker
Best Practices for Sharing Imagery using Amazon Web Services Peter Becker Objectives Making Imagery Accessible Store massive volumes of imagery on inexpensive cloud storage Use elastic compute for image
How To Install An Aneka Cloud On A Windows 7 Computer (For Free)
MANJRASOFT PTY LTD Aneka 3.0 Manjrasoft 5/13/2013 This document describes in detail the steps involved in installing and configuring an Aneka Cloud. It covers the prerequisites for the installation, the
Anwendungsintegration und Workflows mit UNICORE 6
Mitglied der Helmholtz-Gemeinschaft Anwendungsintegration und Workflows mit UNICORE 6 Bernd Schuller und UNICORE-Team Jülich Supercomputing Centre, Forschungszentrum Jülich GmbH 26. November 2009 D-Grid
NIH Commons Overview, Framework & Pilots - Version 1. The NIH Commons
The NIH Commons Summary The Commons is a shared virtual space where scientists can work with the digital objects of biomedical research, i.e. it is a system that will allow investigators to find, manage,
Cloud Computing. Alex Crawford Ben Johnstone
Cloud Computing Alex Crawford Ben Johnstone Overview What is cloud computing? Amazon EC2 Performance Conclusions What is the Cloud? A large cluster of machines o Economies of scale [1] Customers use a
Creating A Galactic Plane Atlas With Amazon Web Services
Creating A Galactic Plane Atlas With Amazon Web Services G. Bruce Berriman 1*, Ewa Deelman 2, John Good 1, Gideon Juve 2, Jamie Kinney 3, Ann Merrihew 3, and Mats Rynge 2 1 Infrared Processing and Analysis
ArcGIS 10.3 Server on Amazon Web Services
ArcGIS 10.3 Server on Amazon Web Services Copyright 1995-2015 Esri. All rights reserved. Table of Contents Introduction What is ArcGIS Server on Amazon Web Services?............................... 5 Quick
Procedia - Social and Behavioral Sciences 141 ( 2014 ) 10 14 WCLTA 2013. Applying Virtualization Technology in Security Education
Available online at www.sciencedirect.com ScienceDirect Procedia - Social and Behavioral Sciences 141 ( 2014 ) 10 14 WCLTA 2013 Applying Virtualization Technology in Security Education Wenjuan Xu a *,
Talari Virtual Appliance CT800. Getting Started Guide
Talari Virtual Appliance CT800 Getting Started Guide March 18, 2015 Table of Contents About This Guide... 2 References... 2 Request for Comments... 2 Requirements... 3 AWS Resources... 3 Software License...
Cloud computing is a marketing term that means different things to different people. In this presentation, we look at the pros and cons of using
Cloud computing is a marketing term that means different things to different people. In this presentation, we look at the pros and cons of using Amazon Web Services rather than setting up a physical server
A Study on Analysis and Implementation of a Cloud Computing Framework for Multimedia Convergence Services
A Study on Analysis and Implementation of a Cloud Computing Framework for Multimedia Convergence Services Ronnie D. Caytiles and Byungjoo Park * Department of Multimedia Engineering, Hannam University
Chapter 4 Cloud Computing Applications and Paradigms. Cloud Computing: Theory and Practice. 1
Chapter 4 Cloud Computing Applications and Paradigms Chapter 4 1 Contents Challenges for cloud computing. Existing cloud applications and new opportunities. Architectural styles for cloud applications.
Viswanath Nandigam Sriram Krishnan Chaitan Baru
Viswanath Nandigam Sriram Krishnan Chaitan Baru Traditional Database Implementations for large-scale spatial data Data Partitioning Spatial Extensions Pros and Cons Cloud Computing Introduction Relevance
ArcGIS for Server in the Amazon Cloud. Michele Lundeen Esri
ArcGIS for Server in the Amazon Cloud Michele Lundeen Esri What we will cover ArcGIS for Server in the Amazon Cloud Why How Extras Why do you need ArcGIS Server? Some examples Publish - Dynamic Map Services
CSE 344 Introduction to Data Management. Section 9: AWS, Hadoop, Pig Latin TA: Yi-Shu Wei
CSE 344 Introduction to Data Management Section 9: AWS, Hadoop, Pig Latin TA: Yi-Shu Wei Homework 8 Big Data analysis on billion triple dataset using Amazon Web Service (AWS) Billion Triple Set: contains
Cloud Computing for Education Workshop
Cloud Computing for Education Workshop 2012 Copyright REZA CURTMOLA, NJIT What Have We Learned So Far? We have learned several cloud applications (Dropbox, Google Docs, Piazza, etc.) Dropbox SugarSync
Web Application Deployment in the Cloud Using Amazon Web Services From Infancy to Maturity
P3 InfoTech Solutions Pvt. Ltd http://www.p3infotech.in July 2013 Created by P3 InfoTech Solutions Pvt. Ltd., http://p3infotech.in 1 Web Application Deployment in the Cloud Using Amazon Web Services From
A Data Dependency Based Strategy for Intermediate Data Storage in Scientific Cloud Workflow Systems *
A Data Dependency Based Strategy for Intermediate Data Storage in Scientific Cloud Workflow Systems * Dong Yuan, Yun Yang, Xiao Liu, Gaofeng Zhang, Jinjun Chen Faculty of Information and Communication
FREE computing using Amazon EC2
FREE computing using Amazon EC2 Seong-Hwan Jun 1 1 Department of Statistics Univ of British Columbia Nov 1st, 2012 / Student seminar Outline Basics of servers Amazon EC2 Setup R on an EC2 instance Stat
Developing High-Performance, Scalable, cost effective storage solutions with Intel Cloud Edition Lustre* and Amazon Web Services
Reference Architecture Developing Storage Solutions with Intel Cloud Edition for Lustre* and Amazon Web Services Developing High-Performance, Scalable, cost effective storage solutions with Intel Cloud
On Demand Satellite Image Processing
On Demand Satellite Image Processing Next generation technology for processing Terabytes of imagery on the Cloud WHITEPAPER MARCH 2015 Introduction Profound changes are happening with computing hardware
Cheminformatics in the Cloud. Michael A. Dippolito DeltaSoft, Inc. 3-June-2009 ChemAxon European User Group Meeting
Cheminformatics in the Cloud Michael A. Dippolito DeltaSoft, Inc. 3-June-2009 ChemAxon European User Group Meeting DeltaSoft Specializing in R&D Informatics since 1996 Based in New Jersey, USA Long term
Simulation-based Evaluation of an Intercloud Service Broker
Simulation-based Evaluation of an Intercloud Service Broker Foued Jrad, Jie Tao and Achim Streit Steinbuch Centre for Computing, SCC Karlsruhe Institute of Technology, KIT Karlsruhe, Germany {foued.jrad,
Collaborative & Integrated Network & Systems Management: Management Using Grid Technologies
2011 International Conference on Computer Communication and Management Proc.of CSIT vol.5 (2011) (2011) IACSIT Press, Singapore Collaborative & Integrated Network & Systems Management: Management Using
