SSI-OSCAR Single System Image - Open Source Cluster Application Resources



Similar documents
The Penguin in the Pail OSCAR Cluster Installation Tool

Sun Grid Engine Package for OSCAR A Google SoC 2005 Project

Kerrighed / XtreemOS cluster flavour

Planning the Installation and Installing SQL Server

Distributed Operating Systems. Cluster Systems

MOSIX: High performance Linux farm

Analysis and Implementation of Cluster Computing Using Linux Operating System

Globus Striped GridFTP Framework and Server. Raj Kettimuthu, ANL and U. Chicago

Achieving Performance Isolation with Lightweight Co-Kernels

Kerrighed: use cases. Cyril Brulebois. Kerrighed. Kerlabs

Highly Reliable Linux HPC Clusters: Self-awareness Approach

HDFS Users Guide. Table of contents

RED HAT ENTERPRISE VIRTUALIZATION FOR SERVERS: COMPETITIVE FEATURES

Red Hat enterprise virtualization 3.0 feature comparison

How to Install an OSCAR Cluster Quick Start Install Guide Software Version 4.2 Documentation Version 4.2

Plug-and-play Virtual Appliance Clusters Running Hadoop. Dr. Renato Figueiredo ACIS Lab - University of Florida

Assignment # 1 (Cloud Computing Security)

CATS-i : LINUX CLUSTER ADMINISTRATION TOOLS ON THE INTERNET

Eloquence Training What s new in Eloquence B.08.00

Installing and Configuring Windows Server Module Overview 14/05/2013. Lesson 1: Planning Windows Server 2008 Installation.

Using Red Hat Network Satellite Server to Manage Dell PowerEdge Servers

WebLogic Server Foundation Topology, Configuration and Administration

Is Virtualization Killing SSI Research?

Managing your Red Hat Enterprise Linux guests with RHN Satellite

XTM Web 2.0 Enterprise Architecture Hardware Implementation Guidelines. A.Zydroń 18 April Page 1 of 12

Technical Overview of Windows HPC Server 2008

FOR SERVERS 2.2: FEATURE matrix

Computer Virtualization in Practice

Oracle Solaris: Aktueller Stand und Ausblick

How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (

Intellicus Enterprise Reporting and BI Platform

Veritas Storage Foundation and High Availability Solutions Getting Started Guide

System Software for High Performance Computing. Joe Izraelevitz

CS 377: Operating Systems. Outline. A review of what you ve learned, and how it applies to a real operating system. Lecture 25 - Linux Case Study

RED HAT ENTERPRISE VIRTUALIZATION

GridFTP: A Data Transfer Protocol for the Grid

Virtualization Technologies and Blackboard: The Future of Blackboard Software on Multi-Core Technologies

Cloud Based Application Architectures using Smart Computing

10215A Implementing and Managing Microsoft Server Virtualization

Microsoft SQL Server versus IBM DB2 Comparison Document (ver 1) A detailed Technical Comparison between Microsoft SQL Server and IBM DB2

High Availability Solutions for the MariaDB and MySQL Database

Simplest Scalable Architecture

installation administration and monitoring of beowulf clusters using open source tools

Monitoring Clusters and Grids

IBM Endpoint Manager Version 9.1. Patch Management for Red Hat Enterprise Linux User's Guide

Linux Cluster Computing An Administrator s Perspective

Automated deployment of virtualization-based research models of distributed computer systems

VMware Server 2.0 Essentials. Virtualization Deployment and Management

Petascale Software Challenges. Piyush Chaudhary High Performance Computing

Deploying Business Virtual Appliances on Open Source Cloud Computing

Red Hat Network Satellite Management and automation of your Red Hat Enterprise Linux environment

Cluster Implementation and Management; Scheduling

Implementing Internet Storage Service Using OpenAFS. Sungjin Dongguen Arum

Red Hat Satellite Management and automation of your Red Hat Enterprise Linux environment

Delivering Quality in Software Performance and Scalability Testing

HDFS Cluster Installation Automation for TupleWare

Automating the deployment of FreeBSD & PC-BSD systems. BSDCan Kris Moore PC-BSD / ixsystems kris@pcbsd.org

Introduction 1 Performance on Hosted Server 1. Benchmarks 2. System Requirements 7 Load Balancing 7

Parallels Cloud Server 6.0

Linux clustering. Morris Law, IT Coordinator, Science Faculty, Hong Kong Baptist University

Implementing and Managing Microsoft Server Virtualization

Software Scalability Issues in Large Clusters

Evaluation of Enterprise Data Protection using SEP Software

Data Centers and Cloud Computing

High Availability of the Polarion Server

Virtualization in Linux

Open-Xchange Whitepaper Open-Xchange Server 6 Architecture Overview

CXS Citrix XenServer 6.0 Administration

Online Remote Data Backup for iscsi-based Storage Systems

New Storage System Solutions

Hardware/Software Guidelines

Copyright by Parallels Holdings, Ltd. All rights reserved.

Parallels Cloud Server 6.0

Chapter 3. Hardware and software requirements

Ghost Process: a Sound Basis to Implement Process Duplication, Migration and Checkpoint/Restart in Linux Clusters

IBM License Metric Tool Version Installing with embedded WebSphere Application Server

Part I Courses Syllabus

Asymmetric Active-Active High Availability for High-end Computing

inforouter V8.0 Server & Client Requirements

Microsoft Compute Clusters in High Performance Technical Computing. Björn Tromsdorf, HPC Product Manager, Microsoft Corporation

EMC DOCUMENTUM xplore 1.1 DISASTER RECOVERY USING EMC NETWORKER

HAOSCAR 2.0: an open source HA-enabling framework for mission critical systems

Migration and Building of Data Centers in IBM SoftLayer with the RackWare Management Module

STeP-IN SUMMIT June 18 21, 2013 at Bangalore, INDIA. Performance Testing of an IAAS Cloud Software (A CloudStack Use Case)

Migration and Building of Data Centers in IBM SoftLayer with the RackWare Management Module

CloudLinux is a proven solution for shared hosting providers that:

Technology in Action. Alan Evans Kendall Martin Mary Anne Poatsy. Eleventh Edition. Copyright 2015 Pearson Education, Inc.

<Insert Picture Here> Oracle Cloud Storage. Morana Kobal Butković Principal Sales Consultant Oracle Hrvatska

StarWind iscsi SAN Software Hands- On Review

Transcription:

2006 OSCAR Symposium St. John's, Newfoundland, Canada May 17, 2006 SSI-OSCAR Single System Image - Open Source Cluster Application Resources Geoffroy Vallée, Thomas Naughton and Stephen L. Scott Oak Ridge National Laboratory, Oak Ridge, TN, USA

Tutorial Structure OSCAR Overview Brief background and project overview Highlight core tools leveraged by OSCAR Describe the extensible package system Summary of spin-off projects SSI-OSCAR Presentation of SSI concept Overview of the Kerrighed SSI Overview of SSI-OSCAR Package

OSCAR Project Overview

OSCAR Background Concept first discussed in January 2000 First organizational meeting in April 2000 Cluster assembly is time consuming & repetitive Nice to offer a toolkit to automate First public release in April 2001 Use best practices for HPC clusters Leverage wealth of open source components Target modest size cluster (single network switch) Form umbrella organization to oversee cluster efforts Open Cluster Group (OCG)

Open Cluster Group Informal group formed to make cluster computing more practical for HPC research and development Membership is open, direct by steering committee Research/Academic Industry Current active working groups [HPC]-OSCAR Thin-OSCAR (diskless) HA-OSCAR (high availability) SSI-OSCAR (single system image) SSS-OSCAR (Scalable Systems Software)

OSCAR Core Organizations

What does OSCAR do? Wizard based cluster software installation Operating system Cluster environment Automatically configures cluster components Increases consistency among cluster builds Reduces time to build / install a cluster Reduces need for expertise

Design Goals Reduce overhead for cluster management Keep the interface simple Provide basic operations of cluster software & node administration Enable others to re-use and extend system deployment tool Leverage best practices whenever possible Native package systems Existing distributions Management, system and applications Extensibility for new Software and Projects Modular meta-package system / API OSCAR Packages Keep it simple for package authors Open Source to foster reuse and community participation Fosters spin-offs to reuse OSCAR framework

OSCAR Wizard

Open Source Cluster Application Resources Step 7 Step 8 Done! Step 1 Start Step 6 Step 2 Step 5 Step 4 Step 3

OSCAR Core

OSCAR Components Administration/Configuration SIS, C3, OPIUM, Kernel-Picker & cluster services (dhcp, nfs, ntp,...) Security: Pfilter, OpenSSH HPC Services/Tools Parallel Libs: MPICH, LAM/MPI, PVM OpenPBS/MAUI HDF5 Ganglia, Clumon, [monitoring systems] Other 3 rd party OSCAR Packages Core Infrastructure/Management System Installation Suite (SIS), Cluster Command & Control (C3), Env- Switcher OSCAR DAtabase (ODA), OSCAR Package Downloader (OPD)

System Installation Suite (SIS) Enhancement suite to the SystemImager tool. Adds SystemInstaller and SystemConfigurator SystemInstaller interface to installation, includes a stand-alone GUI Tksis. Allows for description based image creation. SystemImager base tool used to construct & distribute machine images. SystemConfigurator extension that allows for on-the-fly style configurations once the install reaches the node, e.g. /etc/modules.conf.

System Installation Suite (SIS) Used in OSCAR to install nodes partitions disks, formats disks and installs nodes Construct image of compute node on headnode Directory structure of what the node will contain This is a virtual, chroot able environment /var/lib/systemimager/images/oscarimage/etc/ /usr/ Use rsync to copy only differences in files, so can be used for cluster management maintain image and sync nodes to image

C3 Power Tools Command-line interface for cluster system administration and parallel user tools. Parallel execution cexec Execute across a single cluster or multiple clusters at same time Scatter/gather operations cpush / cget Distribute or fetch files for all node(s)/cluster(s) Used throughout OSCAR and as underlying mechanism for tools like OPIUM s useradd enhancements.

C3 Power Tools Example to run hostname on all nodes of default cluster: $ cexec hostname Example to push an RPM to /tmp on the first 3 nodes $ cpush :1-3 helloworld-1.0.i386.rpm /tmp Example to get a file from node1 and nodes 3-6 $ cget :1,3-6 /tmp/results.dat /tmp

Switcher Switcher provides a clean interface to edit environment without directly tweaking.dot files. e.g. PATH, MANPATH, path for mpicc, etc. Edit/Set at both system and user level. Leverages existing Modules system Changes are made to future shells To help with foot injuries while making shell edits Modules already offers facility for current shell manipulation, but no persistent changes.

OSCAR DAtabase (ODA) Used to store OSCAR cluster data Currently uses MySQL as DB engine User and program friendly interface for database access Capability to extend database commands as necessary.

OSCAR Package Downloader (OPD) Tool to download and extract OSCAR Packages. Can be used for timely package updates Packages that are not included, i.e. 3 rd Party Distribute packages with licensing constraints.

OSCAR Packages

OSCAR Packages Simple way to wrap software & configuration Do you offer package Foo-bar version X? Basic Design goals Keep simple for package authors Modular packaging (each self contained) Timely release/updates Leverage RPM + meta file + scripts, tests, docs, Recently extended to better support RPM, Debs, etc. Repositories for downloading via OPD/OPDer

Package Directory Structure All included packages are in $OSCAR_HOME/packages/ directory with OPD acquired in $OSCAR_PACKAGE_HOME config.xml doc/ distro/ RPMS/ scripts/ SRPMS/ - meta file w/ list of files to install - user.tex, license.tex - distro specific binary packages(s) -[deprecated] binary packages(s) - API scripts - source rpm(s)

Example Package C3 Pre-built C3 software in RPMS/ directory, update: place in distro/<dist-abbrev> Userguide & Installation details in doc/ C3 source package in SRPMS/ Generate configuration file, /etc/c3.conf, using scripts/post_clients List metadata and installation files with target location (server/client) in config.xml

OSCAR Summary Framework for cluster management simplifies installation, configuration and operation reduces time/learning curve for cluster build requires: pre-installed headnode w. supported Linux distribution thereafter: wizard guides user thru setup/install of entire cluster Package-based framework Content: Software + Configuration, Tests, Docs Types: Core: SIS, C3, Switcher, ODA, OPD, APItest, Support Libs Non-core: selected & third-party (PVM, LAM/MPI, Toque/Maui,...) Access: repositories accessible via OPD/OPDer

OSCAR flavors

The OSCAR strategy OSCAR is a snap-shot of best-known-methods for building, programming and using clusters of a reasonable size. To bring uniformity to clusters, foster commercial versions of OSCAR, and make clusters more broadly acceptable. Consortium of research, academic & industry members cooperating in the spirit of open source. Commercially supported Value added instantiations of OSCAR Open Source OSCAR with Linux Other OSCAR Flavors HA-OSCAR, Thin- OSCAR, SSS- OSCAR, SSI-OSCAR

NEC Enhanced OSCAR

NEC's OSCAR-Pro OSCAR'06 Keynote by Erich Focht leverage open source tool two approaches for re-uses: fork / join Commercial enhancements integrate additions when applicable feedback and direction based on user needs

High-Availability OSCAR

HA-OSCAR: RAS Management for HPC cluster: Self-Awareness The first known field-grade open source HA Beowulf cluster release Self-configuration Multi-head Beowulf system HA and HPC clustering techniques to enable critical HPC infrastructure Services: Active/Hot Standby Self-healing with 3-5 sec automatic failover time

Diskless OSCAR

Thin-OSCAR First released in 2003 Why diskless disks are problems costs: initial, power, heat, failures Root RAM technique uses ram disks (/dev/ramxx) compressed RAM disk image transferred by network at each boot minimal system in RAM (~20Mb) Root RAM advantages over NFS less network traffic for the os uses ram only in the exact size of files less stress on the server images are accessed read only nodes more independent from the server

Scalable System Software OSCAR

Scalable System Software Problems Computer centers use incompatible, ad hoc set of systems tools Tools are not designed to scale to multi-teraflop systems Duplication of work to try and scale tools System growth vs. Administrator growth Goals Define standard interfaces for system components Create scalable, standardized management tools (Subsequently) reduce costs & improve efficiency at centers Participants DOE Labs: ORNL, ANL, LBNL, PNNL, SNL, LANL, Ames Academics: NCSA, PSC, SDSC Industry: IBM, Cray, Intel, SGI

SSS Project Overview Map out functional areas Schedulers, Job Mangers System Monitors Accounting & User management Checkpoint/Restart Build & Configuration systems Standardize the system interfaces Open forum of universities, labs, industry reps Define component interfaces in XML Develop communication infrastructure

Meta Scheduler Meta Monitor Meta Manager Meta Services Accounting Scheduler System & Job Monitor Node State Manager Standard XML Interfaces Allocation Management authentication communication Service Directory Event Manager Node Configuration & Build Manager Usage Reports Components written in any mixture of C, C++, Java, Perl, and Python can be integrated into the Scalable Systems Software Suite Job Queue Manager Checkpoint / Restart Process Manager Hardware Infrastructure Manager

SSS-OSCAR Components Bamboo Queue/Job Manager BLCR Berkeley Checkpoint/Restart Gold Accounting & Allocation Management System LAM/MPI (w/ BLCR) Checkpoint/Restart enabled MPI MAUI-SSS Job Scheduler SSSLib SSS Communication library Includes: SD, EM, PM, BCM, NSM, NWI Warehouse Distributed System Monitor MPD2 MPI Process Manager

Single System Image OSCAR

Motivation OSCAR benefits Ease cluster installation and management Setup a traditional Beowulf cluster Parallel application runtime/middle-ware (MPI, PVM) Notion of headnode (e.g. central management of the distributed file system) But how? execute legacy applications not designed for clusters (shared memory applications) execute sequential application designed for large servers

Motivations (2) One approach: resource abstraction adapted to application needs Some examples shared memory applications abstraction of the memory (via a software distributed memory) abstraction of the process/thread scheduling sequential application needing out-of-core computation: abstraction of the memory

SSI - Single System Image Global management of distributed resources: memory, disk, CPU, network Create an abstraction of resources Ultimate SSI: vision of an SMP machine one single huge memory a set of local processors a single file system

SSI-OSCAR Combine benefits of both OSCAR and SSIs ease the installation and management of the cluster ease the use of the cluster Integration of Kerrighed SSI at the kernel level extension of the Linux kernel developed in France, IRISA/INRIA in collaboration with EDF and DGA

Single System Image - Implementation Different approached are possible for the implementation of SSIs User level: middle-ware Kernel level: OS Limitations for functionalities and efficiency (e.g. CONDOR) Complex to develop and maintain (e.g. OpenMOSIX, Kerrighed) Hardware level More expensive (e.g. SGI)

SSI at the Kernel Level Implies the implementation of 4 different features global management of the memory global management of processes/threads global management of disks global management of network communications We will detail these features

Global Memory Management Goal: Extend traditional memory management cluster scale Thread 1 Thread 2 x Memory Node 1

Global Memory Management Goal: Extend traditional memory management cluster scale Thread 1 read/write x Memory Thread 2 Node 1

Global Memory Management Goal: Extend traditional memory management cluster scale Thread Thread 1 2 read/write read/write x Memory Node 1

Global Memory Management Goal: Extend traditional memory management cluster scale Thread Thread 1 2 read/write read/write x Memory Node 1 Thread 3 read/write? Memory Node 2

Global Memory Management Software Distributed Shared Memory Thread 1 Thread 2 Thread 3 Software Distributed Shared Memory Memory Memory Node 1 Node 2

Global Memory Management Software Distributed Shared Memory Thread 1 Thread 2 Thread 3 x Software Distributed Shared Memory Memory Memory Node 1 Node 2

Global Memory Management Software Distributed Shared Memory Thread 1 read x Memory Thread 2 Thread 3 Software Distributed Shared Memory Memory Node 1 Node 2

Global Memory Management Software Distributed Shared Memory Thread 1 x Thread 2 read Memory Thread 3 Software Distributed Shared Memory Memory Node 1 Node 2

Global Memory Management Software Distributed Shared Memory Thread 1 x Thread Thread 2 3 read x Software Distributed Shared Memory Memory Memory Node 1 Node 2

Global Memory Management Software Distributed Shared Memory Thread 1 x Thread Thread 2 3 read x Software Distributed Shared Memory Memory Memory Node 1 Node 2

Global Memory Management Software Distributed Shared Memory Thread 1 x Thread Thread 2 3 read x x Software Distributed Shared Memory Memory Memory Node 1 Node 2

Global Memory Management Software Distributed Shared Memory Thread 1 Thread 2 Thread 3 write x x x Memory Software Distributed Shared Memory Memory Node 1 Node 2

Global Memory Management Software Distributed Shared Memory Thread 1 x INV Thread Thread 2 3 write x INV x Software Distributed Shared Memory Memory Memory Node 1 Node 2

Global Memory Management Software Distributed Shared Memory Thread Thread Thread 1 2 3 write INV x x=6 x INV x Memory Software Distributed Shared Memory Memory Node 1 Node 2

Load Balancing: Introduction We want to have a efficient cluster use High application throughput Efficient resource use Find a good repartition of application into the cluster to use in the best way resources Control of processes/threads during the application deployment and execution

Load Balancing (1) What is the cluster state? Node 1 Node 2 Node 3 Node 4

Load Balancing (1) What is the cluster state? Node 1 Node 2 Node 3 Node 4

Load Balancing (1) What is the cluster state? Node 1 Node 2 (2) When schedule processes? Node 3 Node 4

Load Balancing (1) What is the cluster state? Node 1 Node 2 (2) When schedule processes? (3) What process can be scheduled? Node 3 Node 4

Load Balancing (1) What is the cluster state? Node 1 Node 2 (2) When schedule processes? (3) What process can be scheduled? Node 3 Node 4

Load Balancing (1) What is the cluster state? Node 1 Node 2 (2) When schedule processes? Node 3 Node 4 (3) What process can be scheduled? (4) On which node the process can be scheduled?

Load Balancing (1) What is the cluster state? Node 1 Node 2 (2) When schedule processes? Node 3 Node 4 (3) What process can be scheduled? (4) On which node the process can be scheduled?

Load Balancing (1) What is the cluster state? Node 1 Node 2 (2) When schedule processes? Node 3 Node 4 (3) What process can be scheduled? (4) On which node the process can be scheduled?

Kerrighed: an SSI at the Kernel Level

The Kerrighed Project Aims at globally manage all resources: CPU, memory, disk, IPC Developed in France, IRISA/INRIA, in collaboration with EDF and DGA KerLabs has been created to support Kerrighed Part of the European Project XtreemOS

Global Resource Management in Kerrighed Based on the idea of the extension of existing mechanisms limits modifications inside the kernel keep the same interfaces most of the time To explain that, we will detail two examples global file management global process management

File Management in Linux Read File System File cache Disk manager

File Management in Linux Read File System Lookup_page File cache Disk manager

File Management in Linux Read File System Lookup_page File cache Read_page Disk manager

File Management in Linux Read Write File System Lookup_page File cache Read_page Disk manager

File Management in Linux Read Write File System Lookup_page File cache Read_page Write_page Disk manager

File Management and Containers Read Write File System Container Disk Manager

File Management and Containers Read Write File System Lookup_page File Interface Linker Get_Page Grab_Page Container Disk Manager

File Management and Containers Read Write File System Lookup_page File Interface Linker Get_Page Grab_Page Container First_Touch Invalidate_Page File I/O Linker Read_page Write_page Disk Manager

Containers Architecture System Service System Service Interface Linker Interface Linker Container I/O Linker Resource manager Memory

Containers Overview Share object cluster wide and transparently May be used for the implementation distributed file system parallel file system RAID file system software distributed shared memory distributed cache remote memory allocation etc.

Global Process Management We need to be able to extract a process/thread and move it somewhere else (another node, disk, memory) Concept of process extraction / process abstraction process image simple interfaces to manipulate processes cluster wide => concept of ghost process

Inside the Linux Kernel task_struct Pid Mm files Tty Mm_struct mmap mmap_cache Vm_area_struct Vm_start Vm_end Vm_file Vm_start Vm_end Vm_file Memory image Stack Data Text Physical pages p_pptr p_ysptr task_struct Pid File_struct File_fd File Dentry d_inode d_inode Inode Socket i_pipe Socket i_pipe File task_struct Pid fd_array File * d_inode Socket i_pipe NIC

Inside the Linux Kernel task_struct Pid Mm files Tty Mm_struct mmap mmap_cache Vm_area_struct Vm_start Vm_end Vm_file Vm_start Vm_end Vm_file Memory image Stack Data Text Physical pages p_pptr p_ysptr task_struct Pid File_struct File_fd File Dentry d_inode d_inode Inode Socket i_pipe Socket i_pipe File task_struct Pid fd_array File * d_inode Socket i_pipe NIC Process Meta Data Process Data

Inside the Linux Kernel (2) task_struct Pid Mm files Tty Mm_struct mmap mmap_cache Vm_area_struct Vm_start Vm_end Vm_file Vm_start Vm_end Vm_file Memory image Stack Data Text Physical pages p_pptr p_ysptr task_struct Pid File_struct File_fd File Dentry d_inode d_inode Inode Socket i_pipe Socket i_pipe File task_struct Pid fd_array File * d_inode Socket i_pipe NIC

Inside the Linux Kernel (2) task_struct Pid Mm files Tty Mm_struct mmap mmap_cache Vm_area_struct Vm_start Vm_end Vm_file Vm_start Vm_end Vm_file Memory image Stack Data Text Physical pages p_pptr p_ysptr task_struct Pid File_struct File_fd File Dentry d_inode d_inode Inode Socket i_pipe Socket i_pipe File task_struct Pid fd_array File * d_inode Socket i_pipe NIC

Inside the Linux Kernel (2) task_struct Pid Mm files Tty Mm_struct mmap mmap_cache Vm_area_struct Vm_start Vm_end Vm_file Vm_start Vm_end Vm_file p_pptr p_ysptr task_struct Pid File_struct File_fd File Dentry d_inode d_inode Inode Socket i_pipe Socket i_pipe File task_struct Pid fd_array File * d_inode Socket i_pipe NIC

Inside the Linux Kernel (2) task_struct Pid Mm files Tty p_pptr p_ysptr task_struct Pid Mm_struct mmap mmap_cache File_struct File_fd Vm_area_struct Vm_start Vm_end Vm_file Vm_start Vm_end Vm_file File Dentry d_inode d_inode Container Container Inode Socket i_pipe Socket i_pipe File task_struct Pid fd_array File * d_inode Socket i_pipe NIC

Inside the Linux Kernel (2) task_struct Pid Mm files Tty p_pptr p_ysptr Mm_struct mmap mmap_cache Vm_area_struct Vm_start Vm_end Vm_file Vm_start Vm_end Vm_file Container Container File task_struct Pid File_struct File_fd task_struct Pid fd_array File *

Inside the Linux Kernel (2) task_struct Pid Mm files Tty p_pptr p_ysptr Mm_struct mmap mmap_cache Vm_area_struct Vm_start Vm_end Vm_file Vm_start Vm_end Vm_file Container Container task_struct Pid File_struct File Container File_fd task_struct Pid fd_array File *

Inside the Linux Kernel (2) task_struct Pid Mm files Tty p_pptr p_ysptr Mm_struct mmap mmap_cache Vm_area_struct Vm_start Vm_end Vm_file Vm_start Vm_end Vm_file Container Container task_struct Pid File_struct File_fd File Container KerNet Pipe task_struct Pid fd_array File *

Inside the Linux Kernel (2) task_struct Pid Mm files Tty p_pptr p_ysptr Mm_struct mmap mmap_cache Vm_area_struct Vm_start Vm_end Vm_file Vm_start Vm_end Vm_file Container Container task_struct Pid File_struct File_fd File Container KerNet Pipe task_struct Pid fd_array File * KerNet Socket

The SSI-OSCAR Package

SSI-OSCAR Package Creation 4 steps create binary packages: kernel, modules, tools, includes create configuration scripts: pre/post-install integrate OSCAR tests & documentation create the XML config file => OSCAR packages

OPKG Creation Binary Packages Ease the software integration Guarantee coherency regarding your Linux distribution Package list kernel module lib headers tools

OPKG Creation - Scripts Pre-installation script add the Kerrighed into the image based on kernel_picker (OSCAR tool) Post-installation script: Create configuration files for Kerrighed based on the compute node list adapted to your cluster configuration

SSI-OSCAR Installation Process Download, select and configure the package (OSCAR step 0, 1, and 2) Install the package on the headnode (OSCAR step 3) install the Kerrighed kernel Create the image w/ Kerrighed libs, modules, and tools (OSCAR step 4) Run the pre-install script (OSCAR step 4) add the kernel into the image Run the post-install script (OSCAR step 7) create configuration files

Resources SSI-OSCAR websit http://ssi-oscar.gforge.inria.fr/ Kerrighed website http://www.kerrighed.org/ KerLabs website http://www.kerlabs.com/ OSCAR website http://oscar.openclustergroup.org/ OSCAR Research supported by the Mathematics, Information and Computational Sciences Office, Office of Advanced Scientific Computing Research, Office of Science, U. S. Department of Energy, under contract No. DE-AC05-00OR22725 with UT-Battelle, LLC.