Big Data Research at DKRZ
|
|
|
- Samantha McBride
- 10 years ago
- Views:
Transcription
1 Big Data Research at DKRZ Michael Lautenschlager and Colleagues from DKRZ and Scien:fic Compu:ng Research Group Symposium Big Data in Science Karlsruhe October 7th, 2014
2 Big Data in Climate Research Big data is an all- encompassing term for any collec:on of data sets so large and complex that it becomes difficult to process using tradi:onal data processing applica:ons. (Wikipedia) Big data usually includes data sets with sizes beyond the ability of commonly used sorware tools to capture, curate, manage, and process data within a tolerable elapsed :me. Big data "size" is a constantly moving target, as of 2012 ranging from a few dozen terabytes to many petabytes of data. (Snijders, C., Matzat, U., & Reips, U.- D. (2012). Big Data : Big gaps of knowledge in the field of Internet. Interna'onal Journal of Internet Science, 7, 1-5.) Gartner s "3Vs" model for describing big data: increasing volume (amount of data), velocity (speed of data in and out), and variety (range of data types and sources). (Laney, Douglas. "3D Data Management: Controlling Data Volume, Velocity and Variety". Gartner. Retrieved 6 February 2001.) Big Data in climate research: Satellite data from Earth observa:on (2 V) Climate model data (1.5 V) Data from observa:onal networks (e.g. Tsunami and Earth quarks) (3V) 2
3 Outline The German Climate Compu:ng Center (DKRZ) Infrastructure at DKRZ: present and future Climate model data characteris:c DKRZ s HPC infrastructure developments Climate model data produc:on workflow Future Developments 3
4 The German Climate Computing Center (DKRZ) Founded in 1987 as a na:onal ins:tu:on Operated as a non- profit limited company with four sharehoulders Max Planck Society for Research (55%) The City of Hamburg represented by the University of Hamburg (27%) Alfred Wegener Research Ins:tute in Bremerhaven (9%) Helmholtz Center for Research in Geesthacht (9%) 4
5 Mission DKRZ Partner for Climate Research Maximum Compute Performance. SophisBcated Data Management. Competent Service. 5
6 Climate Modelling Support at DKRZ Basic Workflows: Climate Model Development Climate Model Data Production 6
7 HLRE-2: Architecture Current machine: IBM Power cores, 158 TFLOPS, 20 TB main memory 6 PB disk space, 100 PB tape archive space In 2015 replacement by HLRE-3 7
8 HLRE-2 to HLRE-3 Increases every ten years: Processor speed: 500x Disk capacity: 100x Disk speed: 11x (17x with SSDs) 8
9 HLRE-2: Data Archive Focus: climate model data and related observations Content August 2014 HPSS: 36 PB WDCC: 4 PB (fully documented) HPSS: 8 PB/year (HLRE-3: 75 PB/y) WDCC: PB/year (HLRE-3: 8 PB/y) World Data Centre Climate 9
10 Climate Modelling Example 10 CMIP5 Release WS MPI-M/DKRZ Lautenschlager CAS2K13 (Feb. 2012) (Lautenschlager, (DKRZ), Big Data DKRZ) 10.14
11 Data Amounts CMIP3/CMIP5 CMIP3 / IPCC- AR4 (Report 2007) Par:cipa:on: 17 modelling centres with 25 models In total 36 TB model data central at PCMDI and ca. ½ TB in IPCC DDC at WDCC/ DKRZ as reference data CMIP5 / IPCC- AR5 (Report 2013/2014) Par:cipa:on: 29 modelling groups with 61 models Produced data volume: ca. 10 PB with 640 TB from MPI- ESM CMIP5 requested data volume: ca. 2 PB (in CMIP5 data federa:on) Part of WDCC/DKRZ Data volume for IPCC DDC: 1.6 PB (complete quality assurance process) with 60 TB from MPI- ESM Status CMIP5 data federated archive (August 2014): 2.3 PB for data sets stored in 4.3 Mio Files in 23 data nodes CMIP5 data is more than 50 :mes CMIP3 Extrapola:on for CMIP6 data federa:on: Volume: 150 PB Number of files: 280 Mio Files 11
12 Research: More efficient disk usage Files system: Lustre- ZFS lz4 compression increases energy consump:on by max. 1% But provides compression of 30% in average Potentially applicable to HLRE-3 MDS: meta data server, OSS: object storage server, MDT/OST: meta data / object storage target 12
13 ParallelizaBon on HPC: Scalability ObjecBve of ICON - able to conduct production simulations on target grids as large as grid elements (10 4 x 10 4 x 400) - grid spacing finer than 400 m (100 m being the target) - capable of efficiently using a diversity of advanced high performance computing resources - scaling of the model to many tens, possibly hundreds, of thousands of cores
14 Parallelization on HPC: R2B07 GLOBAL 20km no I/O Efficiency 83% Parallel I/O under development
15 DKRZ offers Services for End-to- End Workflows in Climate Modelling Basic Workflows: Climate Model Develeopment Climate Model Data Production Poster: ISC 14, Leipzig, Juni 2014 URL: 15
16 Climate Model Data Production Workflow Separation: Scientific data analysis Long-term Archiving 16
17 Climate Model Data Production Workflow 1) Data Management Plan The data :me line as well as volumes, structures, access paoerns and storage loca:ons have to be defined as accurate as possible. 2) DKRZ Storage Each DKRZ HPC project has to specify and to apply for compute and storage resources on an annual basis. 3) ESGF StandardizaBon Climate data integra:on into ESGF (Earth System Grid Federa:on) requires standardiza:on in order to make data intercomparable in the federa:on. 4) ESGF Services DKRZ offers a number of services to manage, discover and access climate data in the interna:onal Earth System Grid Federa:on (ESGF). 5) LTA DOKU LTA DOKU stands for in house long- term archiving in the DOKU(menta:on) sec:on of DKRZ s tape archive. Focus here is on internal data access from data providers. 6) LTA WDCC LTA WDCC stands for long- term archiving in the World Data Center Climate (WDCC). This service is open for data from DKRZ HPC projects, data from ESGF but also for data from outside DKRZ. These data are fully integrated in the database system of the WDCC. The full set of metadata is provided in order to allow for data interpreta:on even arer ten years or more without contac:ng the data author. Focus here is on inter- disciplinary data access. 7) DataCite Data PublicaBon DataCite is an interna:onal organiza:on which focuses on publica:on of scien:fic data in context with and to be used in the scien:fic literature. ARer passing the DataCite Data Publica:on services these data en::es are irrevocable, has been assigned a DOI (Digital Object Iden:fier) and key metadata including the cita:on reference are registered in the DataCite repository. The DOI allows for transparent data access and the cita:on reference allows for integra:on into the reference list of a scien:fic publica:on. 17
18 Climate Model Data Federation (ESGF) Service 4 ESGF P2P Architecture: Data Node, Index Node, Compute Node, IdP 18
19 Service 4 ESGF Node 19
20 Services LTA: Data ingestion 20
21 LTA: Data access Services Midtier Appl. Server TDS (or the like) LobServer Archive: files Container: Lobs Storage@DKRZ HPSS DB Layer CERA What Where Who When How 21
22 Seamless end- to- end workflow Model data produc:on analysis long- term archiving Model Code op:miza:on Many cores systems, GPUs File systems Compression, parallel I/O, co- design of storage I/O APIs, object store Object Storage OpenStack SwiR Data access Future Big Data Developments Unique Iden:fier (DOI, Handle System) may replace directory structure Network Bandwidth, SLAs, managed networks 22
Big Data Services at DKRZ
Big Data Services at DKRZ Michael Lautenschlager and Colleagues from DKRZ and Scientific Computing Research Group MPI-M Seminar Hamburg, March 31st, 2015 Big Data in Climate Research Big data is an all-encompassing
CMIP5 Data Management CAS2K13
CMIP5 Data Management CAS2K13 08. 12. September 2013, Annecy Michael Lautenschlager (DKRZ) With Contributions from ESGF CMIP5 Core Data Centres PCMDI, BADC and DKRZ Status DKRZ Data Archive HLRE-2 archive
CMIP6 Data Management at DKRZ
CMIP6 Data Management at DKRZ icas2015 Annecy, France on 13 17 September 2015 Michael Lautenschlager Deutsches Klimarechenzentrum (DKRZ) With contributions from ESGF Executive Committee and WGCM Infrastructure
High Performance Computing OpenStack Options. September 22, 2015
High Performance Computing OpenStack PRESENTATION TITLE GOES HERE Options September 22, 2015 Today s Presenters Glyn Bowden, SNIA Cloud Storage Initiative Board HP Helion Professional Services Alex McDonald,
Big Data and the Earth Observation and Climate Modelling Communities: JASMIN and CEMS
Big Data and the Earth Observation and Climate Modelling Communities: JASMIN and CEMS Workshop on the Future of Big Data Management 27-28 June 2013 Philip Kershaw Centre for Environmental Data Archival
The Ophidia framework: toward cloud- based big data analy;cs for escience Sandro Fiore, Giovanni Aloisio, Ian Foster, Dean Williams
The Ophidia framework: toward cloud- based big data analy;cs for escience Sandro Fiore, Giovanni Aloisio, Ian Foster, Dean Williams Sandro Fiore, Ph.D. CMCC Scientific Computing and Operations Division
UNINETT Sigma2 AS: architecture and functionality of the future national data infrastructure
UNINETT Sigma2 AS: architecture and functionality of the future national data infrastructure Authors: A O Jaunsen, G S Dahiya, H A Eide, E Midttun Date: Dec 15, 2015 Summary Uninett Sigma2 provides High
BENCHMARKING V ISUALIZATION TOOL
Copyright 2014 Splunk Inc. BENCHMARKING V ISUALIZATION TOOL J. Green Computer Scien
Cloudian The Storage Evolution to the Cloud.. Cloudian Inc. Pre Sales Engineering
Cloudian The Storage Evolution to the Cloud.. Cloudian Inc. Pre Sales Engineering Agenda Industry Trends Cloud Storage Evolu4on of Storage Architectures Storage Connec4vity redefined S3 Cloud Storage Use
Introduction History Design Blue Gene/Q Job Scheduler Filesystem Power usage Performance Summary Sequoia is a petascale Blue Gene/Q supercomputer Being constructed by IBM for the National Nuclear Security
Data management and archiving
Data management and archiving for COP / GOP / D_PHASE 4 th COPS Workshop 25./26.09. 2006 Stuttgart Claudia Wunram Hannes Thiemann 25.-26.09.06 / 1 Data archive Long term data archive for COPS, GOP and
Overview on Modern Accelerators and Programming Paradigms Ivan Giro7o [email protected]
Overview on Modern Accelerators and Programming Paradigms Ivan Giro7o [email protected] Informa(on & Communica(on Technology Sec(on (ICTS) Interna(onal Centre for Theore(cal Physics (ICTP) Mul(ple Socket
Interactive Data Visualization with Focus on Climate Research
Interactive Data Visualization with Focus on Climate Research Michael Böttinger German Climate Computing Center (DKRZ) 1 Agenda Visualization in HPC Environments Climate System, Climate Models and Climate
NASA s Big Data Challenges in Climate Science
NASA s Big Data Challenges in Climate Science Tsengdar Lee, Ph.D. High-end Computing Program Manager NASA Headquarters Presented at IEEE Big Data 2014 Workshop October 29, 2014 1 2 7-km GEOS-5 Nature Run
The Data Grid: Towards an Architecture for Distributed Management and Analysis of Large Scientific Datasets
The Data Grid: Towards an Architecture for Distributed Management and Analysis of Large Scientific Datasets!! Large data collections appear in many scientific domains like climate studies.!! Users and
Flexible Scalable Hardware independent. Solutions for Long Term Archiving
Flexible Scalable Hardware independent Solutions for Long Term Archiving More than 20 years of experience in archival storage 2 OA HPA 2010 1992 2000 2004 2007 Mainframe Tape Libraries Open System Tape
Deploying a distributed data storage system on the UK National Grid Service using federated SRB
Deploying a distributed data storage system on the UK National Grid Service using federated SRB Manandhar A.S., Kleese K., Berrisford P., Brown G.D. CCLRC e-science Center Abstract As Grid enabled applications
NetApp High-Performance Computing Solution for Lustre: Solution Guide
Technical Report NetApp High-Performance Computing Solution for Lustre: Solution Guide Robert Lai, NetApp August 2012 TR-3997 TABLE OF CONTENTS 1 Introduction... 5 1.1 NetApp HPC Solution for Lustre Introduction...5
NextGen Infrastructure for Big DATA Analytics.
NextGen Infrastructure for Big DATA Analytics. So What is Big Data? Data that exceeds the processing capacity of conven4onal database systems. The data is too big, moves too fast, or doesn t fit the structures
Data Management in the Cloud: Limitations and Opportunities. Annies Ductan
Data Management in the Cloud: Limitations and Opportunities Annies Ductan Discussion Outline: Introduc)on Overview Vision of Cloud Compu8ng Managing Data in The Cloud Cloud Characteris8cs Data Management
Data Requirements from NERSC Requirements Reviews
Data Requirements from NERSC Requirements Reviews Richard Gerber and Katherine Yelick Lawrence Berkeley National Laboratory Summary Department of Energy Scientists represented by the NERSC user community
Linux Clusters Ins.tute: Turning HPC cluster into a Big Data Cluster. A Partnership for an Advanced Compu@ng Environment (PACE) OIT/ART, Georgia Tech
Linux Clusters Ins.tute: Turning HPC cluster into a Big Data Cluster Fang (Cherry) Liu, PhD [email protected] A Partnership for an Advanced Compu@ng Environment (PACE) OIT/ART, Georgia Tech Targets
Polish National Data Storage. Norbert Meyer, Maciej Brzeźniak, Maciej Stroiński PSNC
Polish National Data Storage Norbert Meyer, Maciej Brzeźniak, Maciej Stroiński PSNC Workshop on Big Data and Open Data, Brussels. May 7-8, 2014 Data = value => needs protection! Data is value:! Expensive
A Service for Data-Intensive Computations on Virtual Clusters
A Service for Data-Intensive Computations on Virtual Clusters Executing Preservation Strategies at Scale Rainer Schmidt, Christian Sadilek, and Ross King [email protected] Planets Project Permanent
SR-IOV In High Performance Computing
SR-IOV In High Performance Computing Hoot Thompson & Dan Duffy NASA Goddard Space Flight Center Greenbelt, MD 20771 [email protected] [email protected] www.nccs.nasa.gov Focus on the research side
Distributed File Systems An Overview. Nürnberg, 30.04.2014 Dr. Christian Boehme, GWDG
Distributed File Systems An Overview Nürnberg, 30.04.2014 Dr. Christian Boehme, GWDG Introduction A distributed file system allows shared, file based access without sharing disks History starts in 1960s
Why are data sharing and reuse so difficult?
Why are data sharing and reuse so difficult? Chris9ne L. Borgman Professor and Presiden9al Chair in Informa9on Studies University of California, Los Angeles hmp://chris9neborgman.info and UCLA Knowledge
T a c k l i ng Big Data w i th High-Performance
Worldwide Headquarters: 211 North Union Street, Suite 105, Alexandria, VA 22314, USA P.571.296.8060 F.508.988.7881 www.idc-gi.com T a c k l i ng Big Data w i th High-Performance Computing W H I T E P A
The Development of Cloud Interoperability
NSC- JST Workshop The Development of Cloud Interoperability Weicheng Huang Na7onal Center for High- performance Compu7ng Na7onal Applied Research Laboratories 1 Outline Where are we? Our experiences before
Cloud Compu)ng in Educa)on and Research
Cloud Compu)ng in Educa)on and Research Dr. Wajdi Loua) Sfax University, Tunisia ESPRIT - December 2014 04/12/14 1 Outline Challenges in Educa)on and Research SaaS, PaaS and IaaS for Educa)on and Research
An Open Dynamic Big Data Driven Applica3on System Toolkit
An Open Dynamic Big Data Driven Applica3on System Toolkit Craig C. Douglas University of Wyoming and KAUST This research is supported in part by the Na3onal Science Founda3on and King Abdullah University
IT of SPIM Data Storage and Compression. EMBO Course - August 27th! Jeff Oegema, Peter Steinbach, Oscar Gonzalez
IT of SPIM Data Storage and Compression EMBO Course - August 27th Jeff Oegema, Peter Steinbach, Oscar Gonzalez 1 Talk Outline Introduction and the IT Team SPIM Data Flow Capture, Compression, and the Data
THE SUN STORAGE AND ARCHIVE SOLUTION FOR HPC
THE SUN STORAGE AND ARCHIVE SOLUTION FOR HPC The Right Data, in the Right Place, at the Right Time José Martins Storage Practice Sun Microsystems 1 Agenda Sun s strategy and commitment to the HPC or technical
How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (
TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 7 th CALL (Tier-0) Contributing sites and the corresponding computer systems for this call are: GCS@Jülich, Germany IBM Blue Gene/Q GENCI@CEA, France Bull Bullx
New Storage System Solutions
New Storage System Solutions Craig Prescott Research Computing May 2, 2013 Outline } Existing storage systems } Requirements and Solutions } Lustre } /scratch/lfs } Questions? Existing Storage Systems
Sun Constellation System: The Open Petascale Computing Architecture
CAS2K7 13 September, 2007 Sun Constellation System: The Open Petascale Computing Architecture John Fragalla Senior HPC Technical Specialist Global Systems Practice Sun Microsystems, Inc. 25 Years of Technical
Hadoop. http://hadoop.apache.org/ Sunday, November 25, 12
Hadoop http://hadoop.apache.org/ What Is Apache Hadoop? The Apache Hadoop software library is a framework that allows for the distributed processing of large data sets across clusters of computers using
Data Management in an International Data Grid Project. Timur Chabuk 04/09/2007
Data Management in an International Data Grid Project Timur Chabuk 04/09/2007 Intro LHC opened in 2005 several Petabytes of data per year data created at CERN distributed to Regional Centers all over the
SAM-FS - Advanced Storage Management Solutions for High Performance Computing Environments
SAM-FS - Advanced Storage Management Solutions for High Performance Computing Environments Contact the speaker: Ernst M. Mutke 3400 Canoncita Lane Plano, TX 75023 Phone: (972) 596-8562, Fax: (972) 596-8552
PACE Predictive Analytics Center of Excellence @ San Diego Supercomputer Center, UCSD. Natasha Balac, Ph.D.
PACE Predictive Analytics Center of Excellence @ San Diego Supercomputer Center, UCSD Natasha Balac, Ph.D. Brief History of SDSC 1985-1997: NSF national supercomputer center; managed by General Atomics
Current Status of FEFS for the K computer
Current Status of FEFS for the K computer Shinji Sumimoto Fujitsu Limited Apr.24 2012 LUG2012@Austin Outline RIKEN and Fujitsu are jointly developing the K computer * Development continues with system
SURFsara Data Services
SURFsara Data Services SUPPORTING DATA-INTENSIVE SCIENCES Mark van de Sanden The world of the many Many different users (well organised (international) user communities, research groups, universities,
Apache Hadoop: The Pla/orm for Big Data. Amr Awadallah CTO, Founder, Cloudera, Inc. [email protected], twicer: @awadallah
Apache Hadoop: The Pla/orm for Big Data Amr Awadallah CTO, Founder, Cloudera, Inc. [email protected], twicer: @awadallah 1 The Problems with Current Data Systems BI Reports + Interac7ve Apps RDBMS (aggregated
Secure Hybrid Cloud Infrastructure for Scien5fic Applica5ons
Secure Hybrid Cloud Infrastructure for Scien5fic Applica5ons Project Members: Paula Eerola Miika Komu MaA Kortelainen Tomas Lindén Lirim Osmani Sasu Tarkoma Salman Toor (Presenter) [email protected]
Hybrid Software Architectures for Big Data. [email protected] @hurence http://www.hurence.com
Hybrid Software Architectures for Big Data [email protected] @hurence http://www.hurence.com Headquarters : Grenoble Pure player Expert level consulting Training R&D Big Data X-data hot-line
HPC and Big Data. EPCC The University of Edinburgh. Adrian Jackson Technical Architect [email protected]
HPC and Big Data EPCC The University of Edinburgh Adrian Jackson Technical Architect [email protected] EPCC Facilities Technology Transfer European Projects HPC Research Visitor Programmes Training
NERSC File Systems and How to Use Them
NERSC File Systems and How to Use Them David Turner! NERSC User Services Group! Joint Facilities User Forum on Data- Intensive Computing! June 18, 2014 The compute and storage systems 2014 Hopper: 1.3PF,
Scala Storage Scale-Out Clustered Storage White Paper
White Paper Scala Storage Scale-Out Clustered Storage White Paper Chapter 1 Introduction... 3 Capacity - Explosive Growth of Unstructured Data... 3 Performance - Cluster Computing... 3 Chapter 2 Current
Introduction to DKRZ and to the Visualization Server halo
Introduction to DKRZ and to the Visualization Server halo Michael Böttinger Niklas Röber Deutsches Klimarechenzentrum -------------------------------------------------------------- German Climate Computing
Building a Top500-class Supercomputing Cluster at LNS-BUAP
Building a Top500-class Supercomputing Cluster at LNS-BUAP Dr. José Luis Ricardo Chávez Dr. Humberto Salazar Ibargüen Dr. Enrique Varela Carlos Laboratorio Nacional de Supercómputo Benemérita Universidad
Hadoop & its Usage at Facebook
Hadoop & its Usage at Facebook Dhruba Borthakur Project Lead, Hadoop Distributed File System [email protected] Presented at the The Israeli Association of Grid Technologies July 15, 2009 Outline Architecture
ETERNUS CS High End Unified Data Protection
ETERNUS CS High End Unified Data Protection Optimized Backup and Archiving with ETERNUS CS High End 0 Data Protection Issues addressed by ETERNUS CS HE 60% of data growth p.a. Rising back-up windows Too
Chapter 7. Using Hadoop Cluster and MapReduce
Chapter 7 Using Hadoop Cluster and MapReduce Modeling and Prototyping of RMS for QoS Oriented Grid Page 152 7. Using Hadoop Cluster and MapReduce for Big Data Problems The size of the databases used in
Scaling Your Data to the Cloud
ZBDB Scaling Your Data to the Cloud Technical Overview White Paper POWERED BY Overview ZBDB Zettabyte Database is a new, fully managed data warehouse on the cloud, from SQream Technologies. By building
LONI Provides UNO High Speed Business Con9nuity A=er Katrina. Speakers: Lonnie Leger, LONI Chris Marshall, UNO
LONI Provides UNO High Speed Business Con9nuity A=er Katrina Speakers: Lonnie Leger, LONI Chris Marshall, UNO Presentation Order Business Con:nuity UNO LONI LONI + UNO Q & A Hurricane Katrina Impact LONI
Big Data. The Big Picture. Our flexible and efficient Big Data solu9ons open the door to new opportuni9es and new business areas
Big Data The Big Picture Our flexible and efficient Big Data solu9ons open the door to new opportuni9es and new business areas What is Big Data? Big Data gets its name because that s what it is data that
NetApp Big Content Solutions: Agile Infrastructure for Big Data
White Paper NetApp Big Content Solutions: Agile Infrastructure for Big Data Ingo Fuchs, NetApp April 2012 WP-7161 Executive Summary Enterprises are entering a new era of scale, in which the amount of data
Get More Scalability and Flexibility for Big Data
Solution Overview LexisNexis High-Performance Computing Cluster Systems Platform Get More Scalability and Flexibility for What You Will Learn Modern enterprises are challenged with the need to store and
LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance
11 th International LS-DYNA Users Conference Session # LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance Gilad Shainer 1, Tong Liu 2, Jeff Layton 3, Onur Celebioglu
EOFS Workshop Paris Sept, 2011. Lustre at exascale. Eric Barton. CTO Whamcloud, Inc. [email protected]. 2011 Whamcloud, Inc.
EOFS Workshop Paris Sept, 2011 Lustre at exascale Eric Barton CTO Whamcloud, Inc. [email protected] Agenda Forces at work in exascale I/O Technology drivers I/O requirements Software engineering issues
HPC Update: Engagement Model
HPC Update: Engagement Model MIKE VILDIBILL Director, Strategic Engagements Sun Microsystems [email protected] Our Strategy Building a Comprehensive HPC Portfolio that Delivers Differentiated Customer Value
Clodoaldo Barrera Chief Technical Strategist IBM System Storage. Making a successful transition to Software Defined Storage
Clodoaldo Barrera Chief Technical Strategist IBM System Storage Making a successful transition to Software Defined Storage Open Server Summit Santa Clara Nov 2014 Data at the core of everything Data is
Archival Storage At LANL Past, Present and Future
Archival Storage At LANL Past, Present and Future Danny Cook Los Alamos National Laboratory [email protected] Salishan Conference on High Performance Computing April 24-27 2006 LA-UR-06-0977 Main points of
Hadoop & its Usage at Facebook
Hadoop & its Usage at Facebook Dhruba Borthakur Project Lead, Hadoop Distributed File System [email protected] Presented at the Storage Developer Conference, Santa Clara September 15, 2009 Outline Introduction
U"lizing the SDSC Cloud Storage Service
U"lizing the SDSC Cloud Storage Service PASIG Conference January 13, 2012 Richard L. Moore [email protected] San Diego Supercomputer Center University of California San Diego SAN DIEGO SUPERCOMPUTER CENTER
Workload Characterization and Analysis of Storage and Bandwidth Needs of LEAD Workspace
Workload Characterization and Analysis of Storage and Bandwidth Needs of LEAD Workspace Beth Plale Indiana University [email protected] LEAD TR 001, V3.0 V3.0 dated January 24, 2007 V2.0 dated August
SURFsara HPC Cloud Workshop
SURFsara HPC Cloud Workshop doc.hpccloud.surfsara.nl UvA workshop 2016-01-25 UvA HPC Course Jan 2016 Anatoli Danezi, Markus van Dijk [email protected] Agenda Introduction and Overview (current
Interna'onal Standards Ac'vi'es on Cloud Security EVA KUIPER, CISA CISSP [email protected] HP ENTERPRISE SECURITY SERVICES
Interna'onal Standards Ac'vi'es on Cloud Security EVA KUIPER, CISA CISSP [email protected] HP ENTERPRISE SECURITY SERVICES Agenda Importance of Common Cloud Standards Outline current work undertaken Define
