Computing at the HL-LHC



Similar documents
Next Generation Operating Systems

(Possible) HEP Use Case for NDN. Phil DeMar; Wenji Wu NDNComm (UCLA) Sept. 28, 2015

Tier0 plans and security and backup policy proposals

DSS. High performance storage pools for LHC. Data & Storage Services. Łukasz Janyst. on behalf of the CERN IT-DSS group

Status and Evolution of ATLAS Workload Management System PanDA

Flash Memory Arrays Enabling the Virtualized Data Center. July 2010

New Design and Layout Tips For Processing Multiple Tasks

Building a Volunteer Cloud

US NSF s Scientific Software Innovation Institutes

Performance monitoring at CERN openlab. July 20 th 2012 Andrzej Nowak, CERN openlab

IT of SPIM Data Storage and Compression. EMBO Course - August 27th! Jeff Oegema, Peter Steinbach, Oscar Gonzalez

Virtualization, Grid, Cloud: Integration Paths for Scientific Computing

Datacenter Operating Systems

Capacity Estimation for Linux Workloads

Today: Data Centers & Cloud Computing" Data Centers"

Parallelism and Cloud Computing

ioscale: The Holy Grail for Hyperscale

Data Centers and Cloud Computing

How Solace Message Routers Reduce the Cost of IT Infrastructure

Copyright 2013, Oracle and/or its affiliates. All rights reserved.

Delivering Quality in Software Performance and Scalability Testing

HYPER-CONVERGED INFRASTRUCTURE STRATEGIES

Achieving Nanosecond Latency Between Applications with IPC Shared Memory Messaging

Cloud Based Application Architectures using Smart Computing

Software-defined Storage Architecture for Analytics Computing

Introducing EEMBC Cloud and Big Data Server Benchmarks

Unisys ClearPath Forward Fabric Based Platform to Power the Weather Enterprise

Stream Processing on GPUs Using Distributed Multimedia Middleware

Data storage services at CC-IN2P3

Cloud Computing and Amazon Web Services

Preview of a Novel Architecture for Large Scale Storage

Programming models for heterogeneous computing. Manuel Ujaldón Nvidia CUDA Fellow and A/Prof. Computer Architecture Department University of Malaga

Introduction to Cloud Computing

HTCondor at the RAL Tier-1

The CMS openstack, opportunistic, overlay, online-cluster Cloud (CMSooooCloud)"

Enabling Technologies for Distributed Computing

Why Choose VMware vsphere for Desktop Virtualization? WHITE PAPER

A Close Look at PCI Express SSDs. Shirish Jamthe Director of System Engineering Virident Systems, Inc. August 2011

The Data Quality Monitoring Software for the CMS experiment at the LHC

Data Centers and Cloud Computing. Data Centers. MGHPCC Data Center. Inside a Data Center

Introduction to the NI Real-Time Hypervisor

Introduction to grid technologies, parallel and cloud computing. Alaa Osama Allam Saida Saad Mohamed Mohamed Ibrahim Gaber

Batch and Cloud overview. Andrew McNab University of Manchester GridPP and LHCb

Protect Data... in the Cloud

Windows Server Virtualization An Overview

Parallel Computing: Strategies and Implications. Dori Exterman CTO IncrediBuild.

BENCHMARKING CLOUD DATABASES CASE STUDY on HBASE, HADOOP and CASSANDRA USING YCSB

Power Management in Cloud Computing using Green Algorithm. -Kushal Mehta COP 6087 University of Central Florida

Cluster, Grid, Cloud Concepts

A REVIEW PAPER ON THE HADOOP DISTRIBUTED FILE SYSTEM

TOP FIVE REASONS WHY CUSTOMERS USE EMC AND VMWARE TO VIRTUALIZE ORACLE ENVIRONMENTS

Last time. Data Center as a Computer. Today. Data Center Construction (and management)

Arif Goelmhd Goelammohamed Solutions Hyperconverged Infrastructure: The How-To and Why Now?

Virtuoso and Database Scalability

Data Centers and Cloud Computing. Data Centers

Datacenters and Cloud Computing. Jia Rao Assistant Professor in CS

Testing & Assuring Mobile End User Experience Before Production. Neotys

Xeon+FPGA Platform for the Data Center

Microsoft Private Cloud Fast Track

Enabling Technologies for Distributed and Cloud Computing

Cloud Storage. Parallels. Performance Benchmark Results. White Paper.

Lecture 2 Cloud Computing & Virtualization. Cloud Application Development (SE808, School of Software, Sun Yat-Sen University) Yabo (Arber) Xu

A Deduplication File System & Course Review

Optimizing Storage for Better TCO in Oracle Environments. Part 1: Management INFOSTOR. Executive Brief

Virtualization. Dr. Yingwu Zhu

Emerging Technology for the Next Decade

ATLAS Petascale Data Processing on the Grid: Facilitating Physics Discoveries at the LHC

Amadeus SAS Specialists Prove Fusion iomemory a Superior Analysis Accelerator

DIABLO TECHNOLOGIES MEMORY CHANNEL STORAGE AND VMWARE VIRTUAL SAN : VDI ACCELERATION

Big Data Performance Growth on the Rise

Computing in High- Energy-Physics: How Virtualization meets the Grid

An Oracle White Paper August Oracle VM 3: Server Pool Deployment Planning Considerations for Scalability and Availability

GPU System Architecture. Alan Gray EPCC The University of Edinburgh

Building a Scalable Big Data Infrastructure for Dynamic Workflows

MaxDeploy Ready. Hyper- Converged Virtualization Solution. With SanDisk Fusion iomemory products

RAMCloud and the Low- Latency Datacenter. John Ousterhout Stanford University

Transcription:

Computing at the HL-LHC Predrag Buncic on behalf of the Trigger/DAQ/Offline/Computing Preparatory Group ALICE: Pierre Vande Vyvre, Thorsten Kollegger, Predrag Buncic; ATLAS: David Rousseau, Benedetto Gorini, Nikos Konstantinidis; CMS: Wesley Smith, Christoph Schwick, Ian Fisk, Peter Elmer ; LHCb: Renaud Legac, Niko Neufeld LHCb Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 1

Computing @ HL-HLC CPU needs (per event) will grow with track multiplicity (pileup) Storage needs are proportional to accumulated luminosity Run1 Run2 Run3 ALICE + LHCb Run4 ATLAS + CMS 1. Resource estimates 2. The probable evolution of the distributed computed technologies Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 2

LHCb & ALICE @ Run 3 40 MHz 40 MHz 50 khz 5-40 MHz Reconstruction + Compression 20 khz (0.1 MB/event) Storage 50 khz (1.5 MB/event) 2 GB/s PEAK OUTPUT 75 GB/s Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 3

ATLAS & CMS @ Run 4 Level 1 Level 1 HLT Storage 5-10 khz (2MB/event) HLT Storage 10 khz (4MB/event) 10-20 GB/s PEAK OUTPUT 40 GB/s Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 4

access rate Big Data standard data processing applicable BIG DATA size Data that exceeds the boundaries and sizes of normal processing capabilities, forcing you to take a non-traditional approach Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 5

How do we score today? Tweeter Stock database Library of Congres Digital collection RAW + Derived data @ Run1 Climatic Data Center database LHC raw data per year YouTube videos per year Digital Health records Google index Facebook new content per year 0 20 40 60 80 100 120 140 160 180 200 PB Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 6

PB Data: Outlook for HL-LHC 450.0 400.0 350.0 300.0 250.0 200.0 150.0 CMS ATLAS ALICE LHCb 100.0 50.0 0.0 Run 1 Run 2 Run 3 Run 4 Very rough estimate of a new RAW data per year of running using a simple extrapolation of current data volume scaled by the output rates. To be added: derived data (ESD, AOD), simulation, user data Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 7

Digital Universe Expansion Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 8

Data storage issues Our data problems may still look small on the scale of storage needs of internet giants Business e-mail, video, music, smartphones, digital cameras generate more and more need for storage The cost of storage will probably continue to go down but Commodity high capacity disks may start to look more like tapes, optimized for multimedia storage, sequential access Need to be combined with flash memory disks for fast random access The residual cost of disk servers will remain While we might be able to write all this data, how long it will take to read it back? Need for sophisticated parallel I/O and processing. + We have to store this amount of data every year and for many years to come (Long Term Data Preservation ) Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 9

Exabyte Scale We are heading towards Exabyte scale Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 10

CPU: Online requirements ALICE & LHCb @ Run 3,4 HLT farms for online reconstruction/compression or event rejections with aim to reduce data volume to manageable levels Event rate x 100, data volume x 10 ALICE estimate: 200k today s core equivalent, 2M HEP-SPEC-06 LHCb estimate: 880k today s core equivalent, 8M HEP-SPEC-06 Looking into alternative platforms to optimize performance and minimize cost GPUs ( 1 GPU = 3 CPUs for ALICE online tracking use case) ARM, Atom ATLAS & CMS @ Run 4 Trigger rate x 10, CMS 4MB/event, ATLAS 2MB/event CMS: Assuming linear scaling with pileup, a total factor of 50 increase (10M HEP-SPEC-06) of HLT power would be needed wrt. today s farm Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 11

MHS06 160 CPU: Online + Offline Moore s law limit 140 120 100 80 GRID ATLAS CMS LHCb ALICE Room for improvement GRID 60 40 Historical growth of 25%/year 20 ONLINE 0 Run 1 Run 2 Run 3 Run 4 Very rough estimate of new CPU requirements for online and offline processing per year of data taking using a simple extrapolation of current requirements scaled by the number of events. Little headroom left, we must work on improving the performance. Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 12

How to improve the performance? Clock frequency Vectors Instruction Pipelining Instruction Level Parallelism (ILP) Hardware threading Multi-core Multi-socket Multi-node } Very little gain to be expected and no action to be taken Potential gain in throughput and in time-to-finish Gain in memory footprint and time-to-finish but not in throughput NEXT TALK Improving the algorithms is the only way to reclaim factors in performance! Running independent jobs per core (as we do now) is optimal solution for High Throughput Computing applications THIS TALK Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 13

WLCG Collaboration Distributed infrastructure of 150 computing centers in 40 countries 300+ k CPU cores (~ 2M HEP-SPEC-06) The biggest site with ~50k CPU cores, 12 T2 with 2-30k CPU cores Distributed data, services and operation infrastructure Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 14

Distributed computing today LHCb Running millions of jobs and processing hundreds of PB of data Efficiency (CPU/Wall time) from 70% (analysis) to 85% (organized activities, reconstruction, simulation) 60-70% of all CPU time on the grid is spent in running simulation Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 15

Can it scale? GlideinWMS LHCb Each experiment today runs their own Grid overlay on top of shared WLCG infrastructure In all cases the underlying distributed computing architecture is similar based on pilot job model Can we scale these systems expected levels? Lots of commonality and potential for consolidation Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 16

Grid Tier model http://cerncourier.com/cws/article/cern/52744 Network capabilities and data access technologies significantly improved our ability to use resources independent of location Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 17

FAX Deployment Global Data Federation B. Bockelman, CHEP 2012 FAX Federated ATLAS Xrootd, AAA Any Data, Any Time, AX Anywhere is a 15PB (CMS), federation, AliEn (ALICE) including ATLAS T3s and multip layers of hierarchy. Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 18

Virtualization Virtualization provides better system utilization by putting all those many cores to work, resulting in power & cost savings. Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 19

Software integration problem Application Libraries Traditional model Horizontal layers Independently developed Maintained by the different groups Different lifecycle Application is deployed on top of the stack Breaks if any layer changes Needs to be certified every time when something changes Results in deployment and support nightmare Difficult to do upgrades Even worse to switch to new OS versions Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 20

Decoupling Apps and Ops Application Libraries Tools Databases OS Application driven approach 1. Start by analysing the application requirements and dependencies 2. Add required tools and libraries Use virtualization to 1. Build minimal OS 2. Bundle all this into Virtual Machine image Separates lifecycles of the application and underlying computing infrastructure Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 21

Cloud: Putting it all together Server consolidation using virtualization improves resource usage API IaaS = Infrastructure as a Service Public, on demand, pay as you go infrastructure with goals and capabilities similar to those of academic grids Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 22

Public Cloud At present 9 large sites/zones up to ~2M CPU cores/site, ~4M total 10 x more cores on 1/10 of the sites compared to our Grid 100 x more users (500,000) Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 23

Could we make our own Cloud? Open Source Cloud components Open Source Cloud middleware VM building tools and infrastructure CernVM+CernVM/FS, boxgrinder.. Common API From the Grid heritage Common Authentication/Authorization High performance global data federation/storage Scheduled file transfer Experiment distributed computing frameworks already adapted to Cloud To do Unified access and cross Cloud scheduling Centralized key services at a few large centers Reduce complexity in terms of number of sites Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 24

Grid on top of Clouds Private CERN Cloud Analysis Reconstruction Custodial Data Store AliEn CAF On Demand Reconstruction Calibration Data Store T1/T2/T3 AliEn O 2 Online-Offline Facility HLT DAQ AliEn CAF On Demand Simulation Analysis Simulation Data Cache Vision Public Cloud(s) Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 25 AliEn

And if this is not enough 18,688 nodes with total of 299,008 processor cores and one GPU per node. And, there are spare CPU cycles available Ideal for event generators/simulation type workloads (60-70% of all CPU cycles used by the experiments). Simulation frameworks must be adapted to efficiently use such resources where available Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 26

Conclusions Resources needed for Computing at HL-LHC are going to be large but not unprecedented Data volume is going to grow dramatically Projected CPU needs are reaching the Moore s law ceiling Grid growth of 25% per year is by far not sufficient Speeding up the simulation is very important Parallelization at all levels is needed for improving performance (application, I/O) Requires reengineering of experiment software New technologies such as clouds and virtualization may help to reduce the complexity and help to fully utilize available resources Predrag Buncic, October 3, 2013 ECFA Workshop Aix-Les-Bains - 27