Lustre SMB Gateway. Integrating Lustre with Windows



Similar documents
Clustered CIFS For Everybody Clustering Samba With CTDB. LinuxTag 2009

Load Balancing and High availability using CTDB + DNS round robin

# Windows Internet Name Serving Support Section: # WINS Support - Tells the NMBD component of Samba to enable its WINS Server ; wins support = no

The Simple High Available Linux File Server. Schlomo Schapiro Principal Consultant Leitung Virtualisierung & Open Source

Large Scale Storage. Orlando Richards, Information Services LCFG Users Day, University of Edinburgh 18 th January 2013

VIDEO SURVEILLANCE WITH SURVEILLUS VMS AND EMC ISILON STORAGE ARRAYS

SMB Direct for SQL Server and Private Cloud

Preparation Guide. How to prepare your environment for an OnApp Cloud v3.0 (beta) deployment.

Step-by-Step Guide to Open-E DSS V7 Active-Active Load Balanced iscsi HA Cluster

Configure Samba with ACL and Active Directory integration Robert LeBlanc BioAg Computer Support, Brigham Young University

BRIDGING EMC ISILON NAS ON IP TO INFINIBAND NETWORKS WITH MELLANOX SWITCHX

New Storage System Solutions

Post-production Video Editing Solution Guide with Microsoft SMB 3 File Serving AssuredSAN 4000

Quantum StorNext. Product Brief: Distributed LAN Client

Introduction to Gluster. Versions 3.0.x

Multiple Public IPs (virtual service IPs) are supported either to cover multiple network segments or to increase network performance.

HPC Update: Engagement Model

Lustre failover experience

NexentaStor Enterprise Backend for CLOUD. Marek Lubinski Marek Lubinski Sr VMware/Storage Engineer, LeaseWeb B.V.

Highly-Available Distributed Storage. UF HPC Center Research Computing University of Florida

RFP-MM Enterprise Storage Addendum 1

Commoditisation of the High-End Research Storage Market with the Dell MD3460 & Intel Enterprise Edition Lustre

Samba. Samba. Samba 2.2.x. Limitations of Samba 2.2.x 1. Interoperating with Windows. Implements Microsoft s SMB protocol

FLOW-3D Performance Benchmark and Profiling. September 2012

Step-by-Step Guide. to configure Open-E DSS V7 Active-Active iscsi Failover on Intel Server Systems R2224GZ4GC4. Software Version: DSS ver. 7.

Comparing SMB Direct 3.0 performance over RoCE, InfiniBand and Ethernet. September 2014

Step-by-Step Guide to Open-E DSS V7 Active-Active iscsi Failover


Scientific Computing Data Management Visions

IT of SPIM Data Storage and Compression. EMBO Course - August 27th! Jeff Oegema, Peter Steinbach, Oscar Gonzalez

SMB Advanced Networking for Fault Tolerance and Performance. Jose Barreto Principal Program Managers Microsoft Corporation

Lessons learned from parallel file system operation

Hyper-V over SMB Remote File Storage support in Windows Server 8 Hyper-V. Jose Barreto Principal Program Manager Microsoft Corporation

Moving Virtual Storage to the Cloud

Cloud Storage. Parallels. Performance Benchmark Results. White Paper.

An Alternative Storage Solution for MapReduce. Eric Lomascolo Director, Solutions Marketing

IOmark- VDI. Nimbus Data Gemini Test Report: VDI a Test Report Date: 6, September

SLIDE 1 Previous Next Exit

GPFS Storage Server. Concepts and Setup in Lemanicus BG/Q system" Christian Clémençon (EPFL-DIT)" " 4 April 2013"

Hadoop on the Gordon Data Intensive Cluster

Springpath Data Platform with Cisco UCS Servers

Globus and the Centralized Research Data Infrastructure at CU Boulder

Investigation of storage options for scientific computing on Grid and Cloud facilities

Hyper-V over SMB: Remote File Storage Support in Windows Server 2012 Hyper-V. Jose Barreto Principal Program Manager Microsoft Corporation

Linux Powered Storage:

Remote PC Guide Series - Volume 1

Linux Administrator (Advance)

Deploying Ceph with High Performance Networks, Architectures and benchmarks for Block Storage Solutions

Moving Virtual Storage to the Cloud. Guidelines for Hosters Who Want to Enhance Their Cloud Offerings with Cloud Storage

IOmark- VDI. HP HP ConvergedSystem 242- HC StoreVirtual Test Report: VDI- HC b Test Report Date: 27, April

LBNC and IBM Corporation Document: LBNC-Install.doc Date: Path: D:\Doc\EPFL\LNBC\LBNC-Install.doc Version: V1.0

Configuration Maximums

Linux Development Environment Description Based on VirtualBox Structure

Transforming the UL into a Big Data University. Current status and planned evolutions

February, 2015 Bill Loewe

High Availability with DRBD & Heartbeat. Chris Barber

Data storage considerations for HTS platforms. George Magklaras -- node manager

Deployment Guide. How to prepare your environment for an OnApp Cloud deployment.

Microsoft SQL Server 2012 on Cisco UCS with iscsi-based Storage Access in VMware ESX Virtualization Environment: Performance Study

Driving IBM BigInsights Performance Over GPFS Using InfiniBand+RDMA

Configuration Maximums

Fujitsu PRIMERGY Servers Portfolio

Object storage in Cloud Computing and Embedded Processing

Four Reasons To Start Working With NFSv4.1 Now

CRIBI. Calcolo Scientifico e Bioinformatica oggi Università di Padova 13 gennaio 2012

Handling Multimedia Under Desktop Virtualization for Knowledge Workers

NERSC File Systems and How to Use Them

IRODS use case : Ciment, the Univ. Grenoble-Alpes HPC center. B.Bzeznik / X.Briand Irods users group meeting 11/06/2015

Syncplicity On-Premise Storage Connector

10th TF-Storage Meeting

NetApp High-Performance Computing Solution for Lustre: Solution Guide

Current Status of FEFS for the K computer

Distributed File System Choices: Red Hat Storage, GFS2 & pnfs

Microsoft Hybrid Cloud IaaS Platforms

Dell Virtual Remote Desktop Reference Architecture. Technical White Paper Version 1.0

Cluster Implementation and Management; Scheduling

Cloud Optimize Your IT

Active Fabric Manager (AFM) Plug-in for VMware vcenter Virtual Distributed Switch (VDS) CLI Guide

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

VMware Virtual SAN Hardware Guidance. TECHNICAL MARKETING DOCUMENTATION v 1.0

How to Choose your Red Hat Enterprise Linux Filesystem

Reference Design: Scalable Object Storage with Seagate Kinetic, Supermicro, and SwiftStack

High Availability Storage

High Availability and Backup Strategies for the Lustre MDS Server

.NET UI Load Balancing & Clustering

HITACHI VIRTUAL STORAGE PLATFORM FAMILY MATRIX

QuickBooks Enterprise Solutions. Linux Database Server Manager Installation and Configuration Guide

PADS GPFS Filesystem: Crash Root Cause Analysis. Computation Institute

BlueArc unified network storage systems 7th TF-Storage Meeting. Scale Bigger, Store Smarter, Accelerate Everything

Performance, Reliability, and Operational Issues for High Performance NAS Storage on Cray Platforms. Cray User Group Meeting June 2007

How to Configure an Initial Installation of the VMware ESXi Hypervisor

HPC Advisory Council

Clustered NAS meets GPFS. Andrew Tridgell LTC ALRT Team

DELL. Dell Microsoft Windows Server 2008 Hyper-V TM Reference Architecture VIRTUALIZATION SOLUTIONS ENGINEERING

Transcription:

Lustre SMB Gateway Integrating Lustre with Windows

Hardware: Old vs New Compute 60 x Dell PowerEdge 1950-8 x 2.6Ghz cores, 16GB, 500GB Sata, 1GBe - Win7 x64 Storage 1 x Dell R510-12 x 2TB Sata, RAID5, 1GBe - Centos 5 Compute 63 x Dell 7910 rack workstation - 24 x 2.5Ghz cores, 64GB, 4 x 900GB 10k SAS, 256GB SSD, 2 x 10Gbe - ESXi 5.5u2 Hypervisor & Win7 VM Storage 4 x Dell R630-8 x 2.4Ghz cores, 64GB, 4 x 600GB 10k SAS, 2 x 10Gbe, 1 x dualport Mellanox ConnectX3-1 x MD3460 Array, 42 x 600GB 10k, 1TB flash - Red Hat EL6 Network Cisco Nexus 2232TM-E Fabric Extender Cisco Nexus 6K 6001 Switches Qlogic 12300 QDR Infiniband switches 2

3

What is CTDB? Clustered implementation of Trivial Database system High Availability service for clustered file-system 4

Why use CTDB? Compute nodes had to be Windows 7 x64 but no native Windows Lustre client existed Save costs on Infiniband network hardware NFS client in Windows is average Opportunity to leverage existing network infrastructure in the Datacentre If Windows could do something well.. CIFS/SMB access would be one of them 5

Differences in CTDB vs SMB CTDB Many hosts for single file-system CTDB service can manage SMB NMB Winbind Host resiliency inbuilt Recovery file lock between CTDB hosts Shared password db SMB Single SMB host per file-system No failover Less potential bandwidth 6

How we implemented CTDB 2 x Physical nodes Simple tdbsam password database Bonded 2 x 10Gbe per host Single QDR link per host to Lustre Local config files [smb.conf, public_addresses,nodes, etc] Shared config / working files [*.tdb, recovery_lock, etc] Round Robin DNS for public IPs 7

HPC Render Cluster Lustre Storage 12 x Object Servers 2 x Metadata Servers QDR Infiniband Network VMware Compute Cluster 63 x Dell 7910 rack workstation 10GBe connectivity per host 1:1 VM to Host mapping CTDB Hosts 4 x Dell R630 2 x 10Gbe in 802.3ad for data 2 x 1Gbe in 802.3ad for heartbeat 1 x Mellanox 2-port ConnectX3 IB ctdb-01 ctdb-02 ctdb-03 ctdb-04 CTDB heartbeat Data Network Networking In rack Cisco Nexus 2232TM-E FEX 4 x 10 Gbit uplink per FEX Pod Cisco Nexus 6001 Switch 6 x 10 Gbit to core network

HPC Render Cluster Lustre Storage 12 x Object Servers 2 x Metadata Servers QDR Infiniband Network VMware Compute Cluster 63 x Dell 7910 rack workstation 10GBe connectivity per host 1:1 VM to Host mapping CTDB Hosts 2 x Dell R630 2 x 10Gbe 802.3ad for data 1 x 1Gbe heartbeat (crossover) 1 x Mellanox 2-port IB HA-SMB Hosts 2 x Dell R630 2 x 10Gbe in 802.3ad for data 1 x 1Gbe heartbeat (crossover) ctdb-01 ctdb-02 HA-smb-01 HA-smb-02 CTDB heartbeat CCS heartbeat Data Network Networking In rack Cisco Nexus 2232TM-E FEX 4 x 10 Gbit uplink per FEX Pod Cisco Nexus 6001 Switch 6 x 10 Gbit to core network

Storage Services Scratch volume Red Hat EL6 host 17 TB Direct attached storage Two node cluster providing an SMB service with HA failover Clustered LVM with EXT4 FS Bandwidth 20Gbit 42 x 600GB DDP, 1TB Flash Archive volume Red Hat EL6 host Samba gateway to 3.0 PB Lustre file-system Clustered TDB (CTDB) used to provide SMB service Lustre FS direct mounted on host CTDB shares a lustre directory Bandwidth up to 40Gbit 10

How the storage is used Storing 8k IMAX film renders Min. 5MB image frame 50,000 frames ~275MB/s stream rate Around 250-300GB of renders for 35 mins footage Streaming from Lustre archive back into final film cuts 11

Sample 8k render 8192 6144 RGB 12

How does CTDB perform in practice? Approximately 30 seconds for complete takeover Tools and documentation are helpful Measured throughput is good ~600MB/s for archive sync from scratch disk over ethernet Normal system load on CTDB hosts with many connections Basic config is reliable 13

What did not work well? Load balancing: TCP connections killed as part of failover Round Robin DNS effect in practice Some notify.sh script handling for monitoring Limited set of default handlers for alerting of cluster state Integrating with existing SMB services: Internal CTDB service scripts modify eth adapter state Splitting out NMB & SMB pids, configs, init scripts was problematic 14

Ideas to improve our CTDB Lustre service Load balancing: Hardware load balancer for initial connection but not transfers Write a software service to manage connections Red Hat EL6 Load Balancer Add-on? Write more comprehensive scripts to alert on cluster state 15

Test results: Lustre disk 2 x CTDB host 12 x clients 4GB data size Sequential read Sequential write Random 512k read Random 512k write 16

Test results: Render Node FEXs 17

Test results: CTDB Node FEXs 18

Conclusions Reasonable performance Reliable and manageable storage service Scalable [as a gateway service] Provides greater accessibility to lustre for non-intensive data applications 19

Basic CTDB Config Mounted Lustre with -o flock [root@rm01 ~]# mount grep lustre 192.168.55.129@o2ib:192.168.55.130@o2ib:/lustre on /lustre type lustre (rw,flock) [root@rm01 ~]#

Basic CTDB Config Created data, and private directory for ctdb operations [root@rm01 ~]# ls -la /lustre/ctdb drwx------ 3 root root 4096 Jan 15 2015.ctdb drwxr-xr-x 3 root root 4096 Feb 3 2015 data [root@rm01 ~]#

Basic CTDB Config Configured /etc/sysconfig/ctdb [root@rm01 ~]# egrep -v '#' /etc/sysconfig/ctdb CTDB_RECOVERY_LOCK=/lustre/ctdb/.ctdb/recovery_lock CTDB_NODES=/etc/ctdb/nodes CTDB_PUBLIC_ADDRESSES=/etc/ctdb/public_addresses CTDB_MANAGES_SAMBA=yes ulimit -n 10000 CTDB_LOGFILE=/var/log/log.ctdb CTDB_DEBUGLEVEL=NOTICE CTDB_NOTIFY_SCRIPT=/etc/ctdb/notify.sh [root@rm01 ~]#

Basic CTDB Config Configured node file (Heartbeat network) [root@rm01 ~]# cat /etc/ctdb/nodes 192.168.100.10 192.168.100.20 [root@rm01 ~]#

Basic CTDB Config Configured public_addresses file (service IP s) [root@rm01 ~]# cat /etc/ctdb/public_addresses 136.186.52.26/24 intelbond0 136.186.52.27/24 intelbond0 [root@rm01 ~]#

Basic CTDB Config Configured public_addresses file (service IP s) [root@rm01 ~]# cat /etc/ctdb/public_addresses 136.186.52.26/24 intelbond0 136.186.52.27/24 intelbond0 [root@rm01 ~]#

Setup smb.conf Basic CTDB Config [root@rm01 ~]# egrep -v '#' /etc/samba/smb.conf [global] workgroup = render server string = Samba Server Version %v netbios name = render-archive interfaces = intelbond0 hosts allow = 127. 136.186.226. 136.186.52. 136.186.12. 136.186.53. clustering = yes ctdbd socket = /var/run/ctdb/ctdbd.socket cluster addresses = 136.186.52.26 136.186.52.27 idmap backend = tdb2 bind interfaces only = no pid directory = /var/run/samba/ctdb private dir = /lustre/ctdb/.ctdb/privdir fileid:mapping = fsname use mmap = no nt acl support = yes ea support = yes log file = /var/log/samba/ctdb/log.%m max log size = 50 security = user passdb backend = tdbsam load printers = no printing = bsd printcap name = /dev/null [archive] comment = Lustre Archive browseable = no writable = yes path = /lustre/ctdb/data/archive [root@rm01 ~]#