Distributed dcache-based storage system of UB RAS



Similar documents
Improvement of computational abilities in computing environments with virtualization technologies

High Availability Databases based on Oracle 10g RAC on Linux

BES-III distributed computing status

IT of SPIM Data Storage and Compression. EMBO Course - August 27th! Jeff Oegema, Peter Steinbach, Oscar Gonzalez

Introduction to Gluster. Versions 3.0.x

The DESY Big-Data Cloud System

Storage database in cloud processing

Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms. Cray User Group Meeting June 2007

Cloud Storage. Parallels. Performance Benchmark Results. White Paper.

Consolidation Technology in the Cloud Data Processing

Processing Large Amounts of Images on Hadoop with OpenCV

Agenda. HPC Software Stack. HPC Post-Processing Visualization. Case Study National Scientific Center. European HPC Benchmark Center Montpellier PSSC

An Affordable Commodity Network Attached Storage Solution for Biological Research Environments.

CMS Tier-3 cluster at NISER. Dr. Tania Moulik

MIGRATING DESKTOP AND ROAMING ACCESS. Migrating Desktop and Roaming Access Whitepaper

An automated system for program parameters fine tuning in the cloud

IBM TSM DISASTER RECOVERY BEST PRACTICES WITH EMC DATA DOMAIN DEDUPLICATION STORAGE

Milestone Solution Partner IT Infrastructure Components Certification Summary

Cluster Computing at HRI

Patrick Fuhrmann. The DESY Storage Cloud

Distributed File System Choices: Red Hat Storage, GFS2 & pnfs

Network Attached Storage. Jinfeng Yang Oct/19/2015

Globus Striped GridFTP Framework and Server. Raj Kettimuthu, ANL and U. Chicago

Storage Virtualization from clusters to grid

Volume Replication INSTALATION GUIDE. Open-E Data Storage Server (DSS )

Performance Analysis of RAIDs in Storage Area Network

Cray DVS: Data Virtualization Service

(Scale Out NAS System)

PARALLELS CLOUD STORAGE

Large Scale Storage. Orlando Richards, Information Services LCFG Users Day, University of Edinburgh 18 th January 2013

Summer Student Project Report

Moving Virtual Storage to the Cloud. Guidelines for Hosters Who Want to Enhance Their Cloud Offerings with Cloud Storage

XFS File System and File Recovery Tools

U-LITE Network Infrastructure

The dcache Storage Element

Virtual Server and Storage Provisioning Service. Service Description

NETWORK ATTACHED STORAGE DIFFERENT FROM TRADITIONAL FILE SERVERS & IMPLEMENTATION OF WINDOWS BASED NAS

High Performance Data-Transfers in Grid Environment using GridFTP over InfiniBand

Running Parameter Sweep applications on Everest cloud platform

SwiftStack Filesystem Gateway Architecture

How To Back Up A Computer To A Backup On A Hard Drive On A Microsoft Macbook (Or Ipad) With A Backup From A Flash Drive To A Flash Memory (Or A Flash) On A Flash (Or Macbook) On

Lessons learned from parallel file system operation

Analisi di un servizio SRM: StoRM

Zadara Storage Cloud A

Moving Virtual Storage to the Cloud

IBM Spectrum Scale vs EMC Isilon for IBM Spectrum Protect Workloads

Scaling Objectivity Database Performance with Panasas Scale-Out NAS Storage

Fibre Channel Overview of the Technology. Early History and Fibre Channel Standards Development

Forschungszentrum Karlsruhe in der Helmholtz-Gemeinschaft. dcache Introduction

HPC Software Requirements to Support an HPC Cluster Supercomputer

THE EXPAND PARALLEL FILE SYSTEM A FILE SYSTEM FOR CLUSTER AND GRID COMPUTING. José Daniel García Sánchez ARCOS Group University Carlos III of Madrid

RFP - Equipment for the Replication of Critical Systems at Bank of Mauritius Tower and at Disaster Recovery Site. 06 March 2014

A High-Performance Storage and Ultra-High-Speed File Transfer Solution

GridFTP GUI: An Easy and Efficient Way to Transfer Data in Grid

Managed GRID or why NFSv4.1 is not enough. Tigran Mkrtchyan for dcache Team

Quantifying the Performance Degradation of IPv6 for TCP in Windows and Linux Networking

Scalable Multi-Node Event Logging System for Ba Bar

Discover Smart Storage Server Solutions

Architecture of distributed network processors: specifics of application in information security systems

RFP-MM Enterprise Storage Addendum 1

Maurice Askinazi Ofer Rind Tony Wong. Cornell Nov. 2, 2010 Storage at BNL

A Comparison on Current Distributed File Systems for Beowulf Clusters

Cluster Scalability of ANSYS FLUENT 12 for a Large Aerodynamics Case on the Darwin Supercomputer

short introduction to linux high availability description of problem and solution possibilities linux tools

How To Build A Clustered Storage Area Network (Csan) From Power All Networks

HPC and Big Data. EPCC The University of Edinburgh. Adrian Jackson Technical Architect

Implementing Network Attached Storage. Ken Fallon Bill Bullers Impactdata

Data storage services at CC-IN2P3

Linux clustering. Morris Law, IT Coordinator, Science Faculty, Hong Kong Baptist University

Panasas at the RCF. Fall 2005 Robert Petkus RHIC/USATLAS Computing Facility Brookhaven National Laboratory. Robert Petkus Panasas at the RCF

Virtualization of a Cluster Batch System

Analysis and Implementation of Cluster Computing Using Linux Operating System

Using High Availability Technologies Lesson 12

An Introduction to ProfitBricks VDC

The PHI solution. Fujitsu Industry Ready Intel XEON-PHI based solution. SC Denver

StorPool Distributed Storage Software Technical Overview

Agenda. Enterprise Application Performance Factors. Current form of Enterprise Applications. Factors to Application Performance.

LSKA 2010 Survey Report Job Scheduler

Traditionally, a typical SAN topology uses fibre channel switch wiring while a typical NAS topology uses TCP/IP protocol over common networking

Shared Parallel File System

STUDY AND SIMULATION OF A DISTRIBUTED REAL-TIME FAULT-TOLERANCE WEB MONITORING SYSTEM

OSG Hadoop is packaged into rpms for SL4, SL5 by Caltech BeStMan, gridftp backend

LBNC and IBM Corporation Document: LBNC-Install.doc Date: Path: D:\Doc\EPFL\LNBC\LBNC-Install.doc Version: V1.0

Ultimate Guide to Oracle Storage

COMPARING STORAGE AREA NETWORKS AND NETWORK ATTACHED STORAGE

HPC performance applications on Virtual Clusters

Deriving semantics from WS-BPEL specifications of parallel business processes on an example

File Transfer Service Monitoring System for LCG

Transcription:

КОМПЬЮТЕРНЫЕ ИССЛЕДОВАНИЯ И МОДЕЛИРОВАНИЕ 2015 Т. 7 3 С. 559 563 СЕКЦИОННЫЕ ДОКЛАДЫ УДК: 004.63 Distributed dcache-based storage system of UB RAS E. Yu. Kuklin 1,3,a, A. V. Sozykin 1,3, A. Yu. Bersenev 1,3, G. F. Masich 2 1 Institute of Mathematics and Mechanics UB RAS, Sofia Kovalevskaya Str. 16, Yekaterinburg, 620990, Russia 2 Institute of Continuous Media Mechanics UB RAS, St. Academ. Koroleva 1, Perm, 614013, Russia 3 Ural Federal University, St. Mira 19, Yekaterinburg, 620002, Russia E-mail: a key@imm.uran.ru Received October 10, 2014 The approach to build territorial distributed storage system for high performance computing environment of UB RAS is presented. The storage system is based on the dcache middleware from the European Middleware Initiative project. The first milestone of distributed storage system implementation includes the data centers at the two UB RAS Regions: Yekaterinburg and Perm. Keywords: network attached storage, parallel NFS, GRID, HPC Распределенная система хранения УРО РАН на основе dcache Е. Ю. Куклин 1,3, А. В. Созыкин 1,3, А. Ю. Берсенёв 1,3, Г. Ф. Масич 2 1 Институт Математики и Механики УрО РАН, Россия, 620990, г. Екатеринбург, ул. Софьи Ковалевской, д. 16 2 Институт Механики Сплошных Сред УрО РАН, Россия, 614013, г. Пермь, ул. Акад. Королёва, д. 1 3 Уральский Федеральный Университет, Россия, 620002, г. Екатеринбург, ул. Мира, д. 19 Представлен подход к созданию территориально-распределенной системы хранения данных для нужд среды высокопроизводительных вычислений УрО РАН. Система основывается на промежуточном программном обеспечении dcache из проекта European Middleware Initiative. Первая очередь реализации системы охватывает вычислительные центры в двух регионах присутствия УрО РАН: г. Екатеринбург и г. Пермь. Ключевые слова: сетевые системы хранения, parallel NFS, ГРИД-технологии, параллельные вычисления Supported by the grant of UB RAS 15-7-1-26 and by the RFBR grant 14-07-96001r_ural_a. Citation: Computer Research and Modeling, 2015, vol. 7, no. 3, pp. 559 563. 2015 Евгений Юрьевич Куклин, Андрей Владимирович Созыкин, Александр Юрьевич Берсенёв, Григорий Фёдорович Масич

560 E. Yu. Kuklin, A. V. Sozykin, A. Yu. Bersenev, G. F. Masich Introduction Works on creating a distributed high-performance computing environment based on GRID technologies are under way at the Ural Branch of the Russian Academy of Sciences. One of the main components of this environment is a distributed data storage system, which aims at integrating storage systems in the Ural regions [Goldshtein et al., 2013]. The system connects various resources, such as computing clusters, supercomputers and experimental setups of Ural institutes and universities. The participants of this storage system project are Institute of Mathematics and Mechanics in Yekaterinburg (computational recourses and storage system) and Institute of Continuous Media Mechanics in Perm (backbone networks). Middleware selection The approach to building territorial distributed storage system based on the dcache [dcache..., 2015] middleware from the European Middleware Initiative (EMI) [EMI..., 2015] project is presented. dcache is a distributed storage system focused on storing large amounts of experimental data. It can run on commodity hardware and allows the construction of storage facilities in hundreds of terabytes, with all the files in it logically organized into a single virtual file system tree. In addition, dcache assumes a simple extension of your storage by adding new nodes and can work with tape libraries. dcache supports a wide range of access protocols. Together with common standard protocols FTP, WebDAV, NFS, grid protocols SRM and GRIDFTP, as well as its own protocol dcap is used. EMI includes three projects for building distributed storage systems: dcache, Disk Pool Manager and Storage Resource Manager. They applied in different projects to build GRID infrastructures, includes WLCG [WLCG..., 2015]. For UB RAS storage dcache was chosen, as it provides support for both GRID and Internet protocols, has a high quality documentation, as well as it is easy to install and administer. An additional reason for choice was the fact that there is a dcache-based store at the Joint Institute for Nuclear Research, and the only one distributed Tier1 center Nordic Data Grid Facility, which storage nodes are located in different Nordic countries [Behrman et al., 2008]. Current stage The first milestone of the implementation of distributed storage system running dcache 2.6 has been currently completed. Figure 1 shows its general plan. For data storing were selected servers by Supermicro. Now there are 4 servers running Scientific Linux 6.5 with usable capacity of 210 TB. We decided to store our data in the XFS file system by Silicon Graphics. It has shown good results in tests for read/write/access to data, and as well as EXT4, is native for Linux and actively developing file system. For example, our test using bonnie++ [Bonnie..., 2015] showed that there is no the distinct advantages between file systems (except random delete, what exactly our system does not imply). So, taking into account the turn of RHEL 7 to XFS, it was decided to leave the XFS. The storage nodes are located in two computing centers: in Institute of Mathematics and Mechanics in Yekaterinburg (3 nodes) and Institute of Continuous Media Mechanics in Perm. The computing centers are separated by a 450 km distance and are joined by a dedicated communication channel. Channel performance provides by DWDM equipment of ECI-Telecom Company. Installed platforms allows transmission of two λ-channels with 10 Gigabit Ethernet technology. Setup and maintenance of backbone network is engaged by Laboratory of Telecommunication and Information Systems in Institute of Continuous Media Mechanics. As shown at the storage system plan, Institute of Mathematics and Mechanics has at its disposal a supercomputer "URAN", which occupies the 9th position in Top50 CIS; Institute of Continuous Me- КОМПЬЮТЕРНЫЕ ИССЛЕДОВАНИЯ И МОДЕЛИРОВАНИЕ

Distributed dcache-based storage system of UB RAS 561 dia Mechanics also has its own cluster. Connection to storage system was performed using the NFS protocol version 4.1 with Parallel NFS [Parallel NFS..., 2015]. On cluster s nodes mounted, as dcache structure suggests, one of the storage system servers (broker host). Parallel NFS allows direct connection between computational and storage nodes for data transfers, removing the traditional NFS-server bottleneck. Fig. 1. General plan of UB RAS DSS There are some unresolved for now drawbacks: how NSF will behave when transmitting data over a long distance, or how to optimize the OS network stack configuration for 10-Gigabit networks (the efficiency of Intel network cards on storage nodes leave much to be desired). Also, there are problems with mounting dcache on servers with a custom build kernel. Network benchmarks First of all, testing has shown that the evident bottleneck is the gigabit network between the storage system and supercomputer "URAN". 10-Gigabit equipment (with switch by Extreme Networks) for the internal network allowed fixing it, and significantly increasing the speed of data exchange. Figure 2 shows the results of the data recording via a dedicated channel between the data centers before and after attaching an additional storage node in Perm. It can be seen that the presence of the local server slightly improves storage system performance. These results were obtained with the IOR benchmark [IOR..., 2015]. Figure 3 displays an attempt to optimize the TCP/IP stack parameters on storage nodes. Our colleagues from Perm conducted tests using iperf [Iperf..., 2015] and various optimization algorithms. As can be seen, the best result was given by bic algorithm, but it could not overcome the threshold of 6 Gbit/s. This is expected has been given by Intel network cards installed on the servers. dcache allows to create multiple copies of files that balance loading between the nodes and optimize the use of storage space. Figure 4 shows that automatic replication does not affect the system performance, so with the hardware RAID arrays, it will increase the reliability of the system, because we do not have at our disposal a tape library for archiving data. 2015, Т. 7, 3, С. 559 563

562 E. Yu. Kuklin, A. V. Sozykin, A. Yu. Bersenev, G. F. Masich Figure 2. Performance recording on DSS from ICMM client with and without an additional storage node in Perm Fig. 3. Comparison of congestion control algorithms for TCP/IP networks Conclusion Now Institute of Mathematics and Mechanics conducted performance and fault tolerance benchmarks of considered storage system. Among them was investigated the effect of additional storage node and replication on the performance of the storage system. Also was selected congestion control algorithm to optimize the network parameters. In the nearest future we plan to pay more attention on security and monitoring. The next stage of implementation will be increasing of storage system КОМПЬЮТЕРНЫЕ ИССЛЕДОВАНИЯ И МОДЕЛИРОВАНИЕ

Distributed dcache-based storage system of UB RAS 563 Fig. 4. Influence of replication on DSS performance capacity and attaching the experimental setups in Institute of Continuous Media Mechanics in Perm and Ural Federal University. Data obtained from them will be recorded to the storage system and processed remotely by supercomputer "URAN", including the ability of process in real time and control the experiments. Further the connection of entire distributed computing environment to international GRID infrastructure Worldwide Large Hadron Collider Computing Grid is planned. We can provide for general use the resources of distributed storage system and computing clusters resources. References Behrman G., Fuhrmann P., Gronager M. and Kleist J. A distributed storage system with dcache // Journal of Physics: Conference Series, 119, 2008; Bonnie++ [online] // 2015. URL: http://www.coker.com.au/bonnie++ (дата обращения: dcache web-page [online] // 2015. URL: http://www.dcache.org (дата обращения: EMI - European Middleware Initiative [online] // 2015. URL: http://www.eu-emi.eu (дата обращения: Goldshtein M.L., Sozykin A.V., Masich G.F., Masich A.G. "Computing resources of UB RAS. Status and prospects." // Parallel Computational Technologies (PCT'2013): proceedings of the international scientific conference. Chelyabinsk, publishing center of SUSU, 2013. P. 330-337; IOR [online] // 2015. URL: http://sourceforge.net/projects/ior-sio (дата обращения: Iperf [online] // 2015. URL: https://iperf.fr (дата обращения: Parallel NFS [online] // 2015. URL: http://www.pnfs.com (дата обращения: 16.01.2015) ; WLCG Worldwide LHC Computing Grid [online] // CERN, Switzerland 2014 URL: http://wlcg.web.cern.ch (дата обращения: 2015, Т. 7, 3, С. 559 563