DoD High Performance Computing Modernization Program: a Distributed Computing Program January Larry Davis
|
|
|
- Gwendoline Glenn
- 10 years ago
- Views:
Transcription
1 DoD High Performance Computing Modernization Program: a Distributed Computing Program January Larry Davis Distribution C: Distribution authorized to U.S. Government Agencies and their contractors for Administrative Use February Other requests for this document shall be referred to Larry Davis, High Performance Computing Modernization Program; Acquisition Sensitive
2 HPC Modernization Program VISION MISSION A pervasive culture existing among DoD s scientists and engineers where they routinely use advanced computational environments to solve the most demanding problems transforming the way DoD does business finding better solutions faster. Accelerate development and transition of advanced defense technologies into superior warfighting capabilities by exploiting and strengthening US leadership in supercomputing, communications, and computational modeling. MB Revised: 11/12/2011 Page-2
3 Technology Transition to the Warfighter HPCMP Simplified Value Chain HPCMP Computational Projects which support RDT&E Acquisition Programs or Operational Information Products US Military The Warfighter MB Revised: 7/26/2011 Page-3
4 5 Execution Priorities Maintain Leadership Continue to effectively deliver high performance computing capabilities to the RDT&E community Pursue new service delivery models that better align the services with unique customer requirements Assume a leadership role in green high performance computing Expand the Base Expand the application of HPC into emerging and non-traditional DoD mission areas Acquisition/engineering communities, cyber, medicine, big data. Increase Customer Productivity Provide portal tools that will support the seamless execution of workflows, make HPC services easily consumable by the next-generation workforce, and address the security barriers associated with connecting distributed desktops to centralized resources Keep pace with available cloud services MB Revised: 09/23/2013 Page-4
5 5 Execution Priorities (continued) Exploit Next-Generation Technologies Anticipate and exploit next-generation HPC technologies Involvement in exascale initiatives, research investments in future HPC capabilities Development of a software investment strategy that will position the DoD to efficiently consume future computing and communications capabilities Develop the Workforce Develop and provide materials and experiences that create both a computationally literate technical workforce for the DoD and a cadre of computational professionals to provision HPC services Investments in STEM programs, support for internships, and involvement in national initiatives that attract and engage with the next-generation workforce MB Revised: 09/23/2013 Page-5
6 HPCMP Organization Chart Regulatory/ Statutory Oversight DASA(R&T) DASD(Research) DASD(C3 & Cyber) Chief Technologist Dr. Roy Campbell Chief Scientist Dr. Robert Wallace ASA(ALT) DASA(R&T) USACE ERDC Program Director Mr. John West Deputy Director Dr. Roy Campbell Special Assistant Dr. Louis Turcotte Advisory Groups User Advocacy Group Computational Technology Area Advisory Panel HPC Advisory Panel HPC Centers Ms. Christine Cuicchi Networking Mr. Ralph McEldowney Software Applications Dr. Andrew Mark CREATE Dr. Douglas Post Business and Contracts Mr. Mickey Robertson Security Ms. Krisa Rowland Workforce Development Ms. Deborah Schwartz Field Activities Army Navy Air Force Agencies Other Plans and Programs Dr. Roy Campbell Dr. Larry Davis Resource Management Dr. Roy Campbell Dr. Larry Davis MB Revised: 9/17/2014 Page-6
7 Service/Agency Approval Authorities (S/AAAs) Allocate Service/Agency controlled CPU hours on HPCMP systems Provide coordination with customer organizations, the HPCMP, and other S/AAAs Ensure Project Leaders and authorized users are performing work in support of DoD Implement requirements surveys, resource allocation, resource monitoring and resource reallocation Provide guidance to users on which HPC assets are appropriate for their projects Trade allocations with other S/AAAs MB Revised: 2/17/2012 Page-7
8 HPCMP Provides a Complete Set of Integrated Advanced Computing Capabilities Developing next-generation scientific and engineering software for advanced computing Fostering similar development by other DoD organizations Advanced computing capabilities via five Centers and an integrated network Efficiently: Costs are 2 5 times less expensive than commercial cloud solutions for similar capability Effectively: HPCMP resource utilization averages approximately 90% of available capacity Fostering the development of a computational workforce Integrating these capabilities to solve complex DoD problems MB Revised: 1/08/2013 Page-8
9 What the HPCMP Offers Currently Approximately a dozen large HPC systems from a variety of vendors at five DoD Supercomputing Resource Centers Connectivity to these centers via a capable wide-area network the Defense Research and Engineering Network Expert help in all aspects of setting up accounts and using HPC systems through the Consolidated Customer Assistance Center and the User Productivity Enhancement, Technology Transfer and Training program Access to many commercial applications software packages for analysis Access to CREATE design modeling tools as they are developed All members of the DoD RDT&E community are eligible to access and use these capabilities at no cost to the end user Page-9
10 Percentage of Total FY15 Requirements by Budget Area 90% Percentage of FY15 HPCMP Requirements 10% Basic Research Applied Research Advanced Tech Development Advanced Component Dev and Prototype System Dev and Demonstration S&T and T&E Management Support (T&E) Operational Systems Development A Recent Thrust in the HPCMP MB Revised: 2/20/2014 Distribution Not Statement yet approved C Applies, for see public cover release page for specifics Page-10
11 Computational Technology Areas (CTAs)* Computational Technology Area Leads Primary DOD Projects Not Primary Computational Structural Mechanics (CSM) Carl Dyka, NSWCD Computational Fluid Dynamics (CFD) Roger Strawn, AMRDEC Computational Chemistry, Biology, and Materials Science (CCM) William Mattson, ARL Computational Electromagnetics and Acoustics (CEA) Andrew Greenwood, AFRL Climate/Weather/Ocean Modeling and Simulation (CWO) William Burnett, NAVO Signal/Image Processing (SIP) Paul Antonik, AFRL Forces Modeling and Simulation (FMS) Jennifer Park, SPAWAR Environmental Quality Modeling and Simulation (EQM) David Richards, ERDC 2 11 Electronics, Networking, and Systems (ENS)/C4I VACANT Integrated Modeling and Test Environments (IMT) VACANT Space and Astrophysical Sciences (SAS) Jill Dahlburg, NRL 8 19 Other Grand Total * FY 2015 requirements survey MB Revised: 2/26/2014 Not yet approved for public release Page-11
12 HPCMP Supercomputing Resources Delivering HPC to the User Navy NRL-MRY ARL Air Force AFRL-RI Navy NUWCN AFRL NSWC-Carderock MHPCC Navy NAWCP Army ARL-MD AFRL-RD AFRL-RD AFRL/SD (2) SSC-SD ERDC NAVY Air Force AEDC Army ATEC (2) SMDC NRL Navy Penn State DTRA DOD Supercomputing Resource Centers (DSRC) Affiliated Resource Centers (ARC) Dedicated HPC Project Investments (DHPI) FY11/12/13 US Air Force Research Lab (AFRL) US Army Research Lab (ARL) Maui HPC Center (MHPCC) NAVY US Army Engineer Research and Development Center (ERDC) US Air Force Research Lab, Information Directorate, AFRL-RI Naval Research Lab (NRL) US Army, Space and Missile Defense Command SMDC US Navy, SSC-SD US Navy, Penn State US Air Force, AFRL/SD (2) DTRA US Army, ATEC (2) US Army, ARL-MD Ÿ US Navy, NRL-MRY Ÿ US Air Force, AEDC US Navy, NSWC-Carderock Ÿ US Navy, NAWCP US Air Force, AFRL/RD US Navy, NUWCN MB Revised: 11/29/2013 Page-12
13 Five Centers Roles and Responsibilities Classified Computing Services Unclassified, but Sensitive, Computing Services CDAAC CDR ARL Unclassified, but Sensitive, Computing Services HPC Helpdesk SW Configuration Management IAM Panel Lead ESM/ECM AFRL Unclassified, but Sensitive Computing Services Open Research Computing Services UDAAC UIT/ezHPC and pie WMS ERDC Unclassified, but Sensitive, Computing Services Classified Computing Services UDR 24x7 Oceanographic Operations Modeling NAVY Unclassified, but Sensitive, Computing Services Special Computing Services UIT Client Development Portal Development MHPCC Strength Through Shared Services & Leadership MB Revised: 11/12/2013 Page-13
14 DOD Supercomputing Resource Centers (DSRCs) Current Capability Location System Available Processors (Cores and Accelerators) Memory (GB) Peak Performance (TeraFLOPS) AFRL SGI ICE X SGI UV Cray XC30 SGI ICE X 73,440 1,004 56,880+(32 GPU) 125,888+(356 Phi)+(356 GPU) 146,880 3, , , ARL Cray XT5 (Classified) IBM idataplex IBM idataplex (Classified) Cray XC40 15,600 20,160 17, ,312+(32 GPU) 41,600 45,696 69, , ERDC Cray XE6 Cray XE6 (Open Research) SGI ICE X 150,912 14, ,440+(32 GPU) 301,824 29, , MHPCC IBM idataplex 12,096 24, NAVY IBM idataplex (Classified) IBM idataplex IBM idataplex Cray XC30 Cray XC30 Cray XC40 Cray XC40 Cray XC40 (Classified) 4,032 19,520 19,520 29,352+(124 Phi) 28,584+(124 Phi)+(32 GPU) 50,208+(168 Phi) 50,208+(168 Phi) 6,720+(24 Phi) 8,960 40,064 40,064 82,752 82, , ,000 27, Total 923,324 2,346,980 26,405 As of: December 2014 FY15 HPC Systems shown in PURPLE (TI-15 systems operational in Q4 FY15) FY15 HPC Systems shown in GREEN (TI-14 systems operational in Q3 FY15) FY14 HPC Systems shown in LIGHT BLUE FY13 HPC Systems shown in ORANGE FY11 HPC Systems shown in BLUE (TI-11/12 HPC systems operational in FY12) FY09 HPC Systems shown in RED Page-14
15 HPC Centers Technical Services Support Technical Services Customer Assistance and Support System and Network Administration System Installation and Integration Application/System Performance Analysis Data Analysis and Assessment Classified and unclassified capability Operators and Operations Support Enterprise System Monitoring (ESM) Outreach MB Revised: 1/18/2013 Page-15
16 Consolidated Customer Assistance Center (CCAC) First point-of-contact for HPCMP user Telephone, user-portal, and support Single ticket tracking system across all sites Tier 0 Self Help/Knowledge Management System on CCAC User Portal Tier 1 and some Tier 2 support - CCAC Tier 2 support - Centers User Ticketing System CCAC Help Desk (Tier 1) Tier 2 CCAC Centers Tiers 2 Vendors 3 Tier 3 support - OEMs, PETTT, & Centers 11,883 tickets processed in FY14 MB Revised: 6/27/2014 Page-16
17 Data Analysis and Assessment Center MB Revised: 1/11/2010 Page-17
18 Baseline Configuration Operating Policies to Make System Diversity User Friendly Common Software Environment: Open-Source Libraries, Utilities &Compilers, Editors, Scripting Tools, Environment Variables, Debuggers, Login Shells, Multi-version SW Modules Standardized Processes: Queue Names, User Notices, Security Access User Support Features: Cross-Center File Transfer, Sample Code Repository, Ticket Life Consistency MB Revised: 1/18/2013 Page-18
19 DOD Supercomputing Resource Centers Highlighted Projects/Services User Dashboard Web display of user-specific metrics and real-time system information Software Configuration Management Sharing software licenses across multiple Centers HPC Portal Desktop access to HPC Advance Reservation Service Reserve CPUs for a specific start time Dedicated Support Partition Dedicated CPUs for an extended period Storage Management Data management for the customer Data Recovery MB Revised: 7/1/2014 Page-19
20 User Dashboard System Status Software Licenses Job Status & History Available Software System Maintenance MB Revised: 7/1/2014 Page-20
21 Software Configuration Management Sharing Licenses Across the Enterprise Package Baseline as of January 2015 Abaqus Analysis tokens: 244 & add-ons; 26 & add-ons (classified) ; Accelrys ANSYS CFD++ Cobalt EnSight Suite GASP Gaussian Suite Site licenses: MS_COMPASS, MS_DISCOVER, MS_DMOL3-SOLID-STATE, MS_CASTEP_PARALLEL; Flexible tokens: 8; MS_VISUALIZER 8; Negotiated WAN model: 23 CFD Solvers, 2048 CFD HPC, 5 Gambit (locked to geographic pools); 1 CFD Solver, 192 CFD HPC (classified) 2 Site licenses: ARL, AFRL; Floating: 2 Serial + 60 Parallel; Add on Modules: EDP & TNE Site licensed at all sites Currently in procurement process 13 Standard, 10 Gold, 5 EnLiten Gold, 2 DR 6 Standard, 2 DR (classified) 7 (unlimited cores per job) 3 site licenses: 1 ARL, 1 AFRL, 1 ERDC LS-DYNA & 3D 500 MATLAB Star-CD, Star-CCM+ TotalView DDT FieldView Pointwise Tecplot 21 (3 Classified), 64 Dist Comp Toolbox, 10 Parallel Comp Toolbox 11 Power Session licenses (unlimited cores per job) 50 architecture-free 8,260 maximum processes 10 Utility Servers, 5 Standard and 5 batch AFRL 5 Utility Servers, 4 AFRL. 4 ARL 10 Utility Servers, 9 AFRL, 8 ARL Utility Server Software INTEL 5 Cluster Toolkit seats MB Revised: 1/8/2013 Page-21
22 HPC Portal Web Supercomputing Quick Access to HPC Easy to Use Similar to a webmail interface No user-installed software or patches Integrated videos, forums, and help Secure Quick DOD CAC-authentication Secured at one server vs. many desktops YubiKey enabled for University and Industry collaborators No desktop install is a security best-practice Powerful Access to more than 50,000 CPU Cores Power-user command-line shell access Many apps at one location Software runs close to large datasets MB Revised: 7/1/2014 Page-22
23 HPC Portal Benefits Applications for DOD-Scale Solutions Productivity via Workflows Integrated analysis and engineering HPC applications CREATE software Domain-specific tools General analysis and visualization Software Developer Kit to build ecosystem Collaboration and Learning Apply the power of supercomputing to impact the mission! MB Revised: 11/12/2013 Page-23
24 HPCMP CREATE Kestrel Delivery Using the HPC Portal Make CREATE analysis and HPC power accessible to the DOD acquisition community from their work locations Guided inputs of Aircraft, Mach, Altitude, Attack Angle, etc. Interactive Plots MB Revised: 11/12/2013 Page-24
25 Number of Responses User Satisfaction Survey FY14 Services, Tools and Capabilities Have you used/taken advantage of any of the following Program services, tools and/or capabilities? (choose all that apply*) % 217 1,017 Responses * % % These 99 respondents moved to the next section % 99 10% 50 0 Archive/Mass Storage 7% ARS 8 ezviz Service % 1% 1% DAAC DSPs PBS HPC Portal SRD ezhpc None of the above 17 2% If only S/AAA role selected, respondent skipped this section. * Allowed for doubled counting. Page-25
26 mod_jk Portal Architecture Apache HTTPD Gateway Server Compute Node Utility Server Tomcat Portal Server Portal Framework Servlets Liferay Portal Compute Node SSH Supercomputers Edge Service Node Login Nodes Login Nodes $WORKDIR Compute Pool Login 0001 Login 0002 Matlab Compiler Compute Node Virtual Apps Virtual Compute Apps Compute Node Nodes VirApps Vi Vir Apps Compute Virtual Compute Node Node Apps Graphics Node Nodes Edge Service Node on HPC SSH Daemon Connection to US Tomcat Node Provides Service User ( tomcat ) Manages sudo for Service User to Perform Actions on Behalf of Users Compute, Graphics, Memory Node (PBS Pool) CWFS ($CENTER) DSRC DREN Internet Page-26
27 Advance Reservation Service ( Advance Reservation Service (ARS) capability is available on all production systems ARS allows a user to reserve a block of time for their use on any system for which the user has an account A reservation can be made up to 24-hours in advance of the requested time slot for up to 8 wall-clock hours of usage A reservation can be made up to 3 months in advance for a block of up to 168 wall-clock hours A single user can reserve up to 50% of ARS resources available on a system Typically, 25% of a system is available for ARS MB Revised: 11/12/2013 Page-27
28 Dedicated Support Partition (DSP) Intended to support large computational work that would benefit from a dedicated system partition and not be addressable with advance reservation service or utility server Long-term recurring work ( days) Developmental testing AFRL DSRC 57K Xeon Ivy Bridge Cray Cluster, 2 DSPs, partition sizes 512 and 5000 cores 73K Xeon Sandy Bridge SGI Cluster, 1 DSP, partition size 1088 cores ARL DSRC 20K Xeon Sandy Bridge IBM Cluster, 1 DSP, partition size 1040 cores 17K Xeon Sandy Bridge IBM Cluster, 1 DSP, partition size 1088 cores ERDC DSRC 150K AMD Interlagos Cray Cluster, 1 DSP, partition size 1184 cores MHPCC DSRC 12K Xeon Sandy Bridge IBM Cluster 6 DSPs, with partition sizes ranging from 32 cores to 1024 cores Navy DSRC 19K Xeon Sandy Bridge IBM Cluster, 1 DSP, partition size 1024 cores MB Revised: 11/12/2013 Page-28
29 Terabytes of Data Stored (Single Copy) Long-Term Data Storage Data Stored at UDR at the End of Each Fiscal Year 90,000 80,000 70,000 Unclassified Storage by HPC Center AFRL 14% 60,000 50,000 40,000 30,000 20,000 10,000 NAVY 42% ARL 6% 0 UDR Growth Moore's Law MHPCC 1% ERDC 37% MB Revised: 9/24/2013 Page-29
30 Dedicated HPC Project Investments (DHPIs) Dedicated HPC Project Investments (DHPIs) are modest-sized HPC systems awarded to technically-sound, mission-critical projects that cannot be performed with shared HPCMP resources due to classification level, special operational requirements (e.g., real-time turnaround) and/or use of emerging technologies Awarded DHPI Proposals FY14 ($4.1M) Real-Time Hadoop/Hbase High Performance Computing System ($0.35M) Army/ATEC, WSMR, NM and Aberdeen Test Center, MD Applications of the Digital Thread to Capability Planning & Analysis ($2.25M) Air Force/AF Lifecycle Management Center, Wright-Patterson AFB, OH On-Demand GPU Applications ($1.5M) Navy/Naval Research Laboratory, DC 9 proposals received: 3 selected, 1 recommended for DSRC support TI BOA available for system acquisition, or obligate by 30 Sept MB Revised: 4/18/2014 Page-30
31 Networking Defense Research & Engineering Network (DREN) DoD s premier RDT&E network focused on S&T and T&E provides separation from Warfighting networks High-bandwidth, low-latency, full-service network Connects DoD Supercomputing Resource Centers (DSRCs) and users Secret-level network overlay on DREN backbone (SDREN) DREN III provides 50 Mbps to 10 Gbps service to DoD sites; across a 100 Gbps backbone Fully supports IPv6 & Multicast Enables research and test events requiring low-jitter, lowlatency and high-throughput One component of the Global Information Grid Sister network to NIPRNet/SIPRNet 0 Mbps Network # of Sites Mbps Mbps 500 1,000 Mbps 1,000-2,500 Mbps DREN ,500+ Mbps SDREN Mbps 10 Gbps 100 Gbps Site Speeds Backbone MB Revised: 2/05/2014 Page-31
32 DREN III Highlights DREN III contract successfully awarded in late 2012 CenturyLink received a 10 year IDIQ award, worth up to $750M New technical capabilities such as Ethernet and Optical Services Significantly increased capability Attribute DREN II DREN III Ratio Backbone (in Gbps) ~10 Aggregate On-ramp BW (in Gbps) ~5 Internet BW (in Gbps) 4 20 ~5 Transition of 149 sites completed by 17June 2014 MB Revised: 2/05/2014 Page-32
33 Resource Management Requirements & Allocations Mission To ensure HPCMP computational resources are appropriately distributed across the DOD RDT&E communities based upon documented requirements and Service priorities Goals Gather, analyze, store, and report requirements, allocation and usage information on DOD computational projects and HPC systems performance to support DOD/Services/Agencies allocation, program planning, acquisition, and operational processes Execute and report customer community special interest activities, such as requirements site visits and Challenge and Frontier Project allocation processes MB Revised: 11/29/2011 Distribution Not Statement yet approved C Applies, for see public cover release page for specifics Requirements Requirements Gathering and Site Visits pie Portal to the Information Environment Allocations Challenge Projects, Frontier Projects, Capability Applications Projects (CAPs) Metrics HPC Systems Performance Page-33
34 Agencies HPCMP Resource Management Process Key Elements Requirements Gathering Project Leaders Principal Investigators S/AAAs Requirements Validation Service/Agency Principals Frontier/Challenge/High- Priority Project Allocation (35%) Army 30% Navy 30% Air Force 30% Resource Allocation Service/Agency Allocations (65%) 10% HPC Utilization Five Centers ARL AFRL ERDC MHPCC NAVY User Satisfaction Survey Site Visits ARCs (4) MB Revised: 2/20/2014 Distribution Not Statement yet approved C Applies, for see public cover release page for specifics Page-34
35 The Information Environment (IE) (and Its Portal (pie)) Authoritative database for all user, project, and subproject information All accounts must be approved by the Service organization s Service/Agency Approval Authority (S/AAA) Requirements questionnaire incorporated into IE All training requirements part of IE Information Assurance tracked now Additional training will be tracked with January release Constantly evolving to accommodate new Program initiatives International Access forms Unix Groups New Software Request Forms Dedicated Support Partitions On-demand downloads available to Center Site Account Managers MB Revised: 2/20/2014 Page-35
36 HPCMP Allocation Management Process Allocate 100% of all resources at DoD Supercomputing Resource Centers Approximately 70% allocated to Services/Agencies for distribution Resources are delegated to the Services/Agencies for allocation to Service/Agency mission priorities 30/30/30/10 (Army/Navy/Air Force/Agencies) split Mechanisms in place to support Service/Agency allocations for high priority, advance reservations, and dedicated workloads Frontier plus Challenge plus Service/Agency High- Priority is targeted at 35% of all resources allocated Challenge Projects will end in FY 2015 FY 2016 target is for Service/Agency High-Priority and Frontier to share the 35% pool Page-36
37 HPC Systems Performance Metrics HPC Systems Performance Metrics are aggregated across all Program systems and sent to the HPC community monthly High-level metrics include: System usage Turnaround time (expansion factor) Value metric combines system usage and turnaround time into one overall metric which characterizes the total value of the computational time to the user community New metrics being developed to track use and efficiency of advance reservation service MB Revised: 12/21/2010 Page-37
38 Tactical Cloudlets Advanced Computing Mobile ad hoc Networking Tactical Cloudlets Power Availability Capacity Ease of Use Portability Hand-Held Field Programmable Gate Array (FPGA) ARM Traditional Desktop/Workstation Multi/Many-core (von Neumann) Non-optimal HPC node Specialized Desktop/Workstation Streaming cores Graphics Processing Unit (GPU) Large-Scale HPC Cluster technology Data and task parallel Optimal Focus on performance tailored per core No wasted processing cycles Power-aware and powerefficient Packing High Performance Computing (HPC) in small footprints Heterogeneous devices (and links) Hostile environments (short term, mobile, limited bandwidth) Cloud services too costly Virtually impossible to acquire real-time processing (via HPC) Efficiently distributed HPC nodes (seeding) Balanced processing capacity with demand Limiting hop distance between client-server
HPCMP New Users Guide Who Are We? Distribution A: Approved for Public release; distribution is unlimited.
HPCMP New Users Guide Who Are We? HPCMP Vision and Mission VISION Our vision is one in which a pervasive culture exists within the DOD that drives the routine use of advanced computational environments
Big Data in Test and Evaluation by Udaya Ranawake (HPCMP PETTT/Engility Corporation)
Big Data in Test and Evaluation by Udaya Ranawake (HPCMP PETTT/Engility Corporation) Approved for Public Release. Distribution Unlimited. Data Intensive Applications in T&E Win-T at ATC Automotive Data
UNCLASSIFIED. R-1 Program Element (Number/Name) PE 0603461A / High Performance Computing Modernization Program. Prior Years FY 2013 FY 2014 FY 2015
Exhibit R-2, RDT&E Budget Item Justification: PB 2015 Army Date: March 2014 2040: Research, Development, Test & Evaluation, Army / BA 3: Advanced Technology Development (ATD) COST ($ in Millions) Prior
High Performance Computing Modernization Program Mass Storage
High Performance Computing Modernization Program Mass Storage September 21, 2012 Designing Storage Architectures for Digital Collections Library of Congress Gene McGill, [email protected] Arctic Region Supercomputing
Condor Supercomputer
Introducing the Air Force Research Laboratory Information Directorate Condor Supercomputer DOD s Largest Interactive Supercomputer Ribbon Cutting Ceremony AFRL, Rome NY 1 Dec 2010 Approved for Public Release
Dr. Raju Namburu Computational Sciences Campaign U.S. Army Research Laboratory. The Nation s Premier Laboratory for Land Forces UNCLASSIFIED
Dr. Raju Namburu Computational Sciences Campaign U.S. Army Research Laboratory 21 st Century Research Continuum Theory Theory embodied in computation Hypotheses tested through experiment SCIENTIFIC METHODS
Data Intensive Science and Computing
DEFENSE LABORATORIES ACADEMIA TRANSFORMATIVE SCIENCE Efficient, effective and agile research system INDUSTRY Data Intensive Science and Computing Advanced Computing & Computational Sciences Division University
UNCLASSIFIED R-1 ITEM NOMENCLATURE FY 2013 OCO
Exhibit R-2, RDT&E Budget Item Justification: PB 2013 Office of Secretary Of Defense DATE: February 2012 COST ($ in Millions) FY 2011 FY 2012 Base PE 0603755D8Z: High Performance Computing OCO Total FY
Enterprise HPC & Cloud Computing for Engineering Simulation. Barbara Hutchings Director, Strategic Partnerships ANSYS, Inc.
Enterprise HPC & Cloud Computing for Engineering Simulation Barbara Hutchings Director, Strategic Partnerships ANSYS, Inc. Historical Perspective Evolution of Computing for Simulation Pendulum swing: Centralized
SGI HPC Systems Help Fuel Manufacturing Rebirth
SGI HPC Systems Help Fuel Manufacturing Rebirth Created by T A B L E O F C O N T E N T S 1.0 Introduction 1 2.0 Ongoing Challenges 1 3.0 Meeting the Challenge 2 4.0 SGI Solution Environment and CAE Applications
Mississippi State University High Performance Computing Collaboratory Brief Overview. Trey Breckenridge Director, HPC
Mississippi State University High Performance Computing Collaboratory Brief Overview Trey Breckenridge Director, HPC Mississippi State University Public university (Land Grant) founded in 1878 Traditional
The Maui High Performance Computing Center Department of Defense Supercomputing Resource Center (MHPCC DSRC) Hadoop Implementation on Riptide - -
The Maui High Performance Computing Center Department of Defense Supercomputing Resource Center (MHPCC DSRC) Hadoop Implementation on Riptide - - Hadoop Implementation on Riptide 2 Table of Contents Executive
Outline. High Performance Computing (HPC) Big Data meets HPC. Case Studies: Some facts about Big Data Technologies HPC and Big Data converging
Outline High Performance Computing (HPC) Towards exascale computing: a brief history Challenges in the exascale era Big Data meets HPC Some facts about Big Data Technologies HPC and Big Data converging
HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK
HETEROGENEOUS HPC, ARCHITECTURE OPTIMIZATION, AND NVLINK Steve Oberlin CTO, Accelerated Computing US to Build Two Flagship Supercomputers SUMMIT SIERRA Partnership for Science 100-300 PFLOPS Peak Performance
:Introducing Star-P. The Open Platform for Parallel Application Development. Yoel Jacobsen E&M Computing LTD [email protected]
:Introducing Star-P The Open Platform for Parallel Application Development Yoel Jacobsen E&M Computing LTD [email protected] The case for VHLLs Functional / applicative / very high-level languages allow
Overview of HPC Resources at Vanderbilt
Overview of HPC Resources at Vanderbilt Will French Senior Application Developer and Research Computing Liaison Advanced Computing Center for Research and Education June 10, 2015 2 Computing Resources
GPU System Architecture. Alan Gray EPCC The University of Edinburgh
GPU System Architecture EPCC The University of Edinburgh Outline Why do we want/need accelerators such as GPUs? GPU-CPU comparison Architectural reasons for GPU performance advantages GPU accelerated systems
Bringing Big Data Modelling into the Hands of Domain Experts
Bringing Big Data Modelling into the Hands of Domain Experts David Willingham Senior Application Engineer MathWorks [email protected] 2015 The MathWorks, Inc. 1 Data is the sword of the
Maui High Performance Computing Center Overview
Maui High Performance Computing Center Overview March 2013 Peter Colvin MHPCC Director of Outreach & Program Development University of Hawaii 1 Agenda Background Strategy Resources Clients Summary 2 Background
HPC technology and future architecture
HPC technology and future architecture Visual Analysis for Extremely Large-Scale Scientific Computing KGT2 Internal Meeting INRIA France Benoit Lange [email protected] Toàn Nguyên [email protected]
Cloud Computing through Virtualization and HPC technologies
Cloud Computing through Virtualization and HPC technologies William Lu, Ph.D. 1 Agenda Cloud Computing & HPC A Case of HPC Implementation Application Performance in VM Summary 2 Cloud Computing & HPC HPC
IBM Platform Computing Cloud Service Ready to use Platform LSF & Symphony clusters in the SoftLayer cloud
IBM Platform Computing Cloud Service Ready to use Platform LSF & Symphony clusters in the SoftLayer cloud February 25, 2014 1 Agenda v Mapping clients needs to cloud technologies v Addressing your pain
COMP/CS 605: Intro to Parallel Computing Lecture 01: Parallel Computing Overview (Part 1)
COMP/CS 605: Intro to Parallel Computing Lecture 01: Parallel Computing Overview (Part 1) Mary Thomas Department of Computer Science Computational Science Research Center (CSRC) San Diego State University
Accelerating CFD using OpenFOAM with GPUs
Accelerating CFD using OpenFOAM with GPUs Authors: Saeed Iqbal and Kevin Tubbs The OpenFOAM CFD Toolbox is a free, open source CFD software package produced by OpenCFD Ltd. Its user base represents a wide
How To Build A Supermicro Computer With A 32 Core Power Core (Powerpc) And A 32-Core (Powerpc) (Powerpowerpter) (I386) (Amd) (Microcore) (Supermicro) (
TECHNICAL GUIDELINES FOR APPLICANTS TO PRACE 7 th CALL (Tier-0) Contributing sites and the corresponding computer systems for this call are: GCS@Jülich, Germany IBM Blue Gene/Q GENCI@CEA, France Bull Bullx
Enabling the SmartGrid through Cloud Computing
Enabling the SmartGrid through Cloud Computing April 2012 Creating Value, Delivering Results 2012 eglobaltech Incorporated. Tech, Inc. All rights reserved. 1 Overall Objective To deliver electricity from
Performance Analysis of Web based Applications on Single and Multi Core Servers
Performance Analysis of Web based Applications on Single and Multi Core Servers Gitika Khare, Diptikant Pathy, Alpana Rajan, Alok Jain, Anil Rawat Raja Ramanna Centre for Advanced Technology Department
Estimating Cloud Computing Costs: Practical Questions for Programs
Estimating Cloud Computing Costs: Practical Questions for Programs Kathryn Connor RAND Corporation June 2014 In 2010, the OMB proposed Federal Agencies consolidate data centers to save money. In 2013,
Compute Canada Technology Briefing
Compute Canada Technology Briefing November 12, 2015 Introduction Compute Canada, in partnership with regional organizations ACENET, Calcul Québec, Compute Ontario and WestGrid, leads the acceleration
HPC Wales Skills Academy Course Catalogue 2015
HPC Wales Skills Academy Course Catalogue 2015 Overview The HPC Wales Skills Academy provides a variety of courses and workshops aimed at building skills in High Performance Computing (HPC). Our courses
Simulation Platform Overview
Simulation Platform Overview Build, compute, and analyze simulations on demand www.rescale.com CASE STUDIES Companies in the aerospace and automotive industries use Rescale to run faster simulations Aerospace
XSEDE Service Provider Software and Services Baseline. September 24, 2015 Version 1.2
XSEDE Service Provider Software and Services Baseline September 24, 2015 Version 1.2 i TABLE OF CONTENTS XSEDE Production Baseline: Service Provider Software and Services... i A. Document History... A-
The Top Six Advantages of CUDA-Ready Clusters. Ian Lumb Bright Evangelist
The Top Six Advantages of CUDA-Ready Clusters Ian Lumb Bright Evangelist GTC Express Webinar January 21, 2015 We scientists are time-constrained, said Dr. Yamanaka. Our priority is our research, not managing
Red Hat Enterprise Linux 6. Stanislav Polášek ELOS Technologies [email protected]
Stanislav Polášek ELOS Technologies [email protected] Red Hat - an Established Global Leader Compiler Development Identity & Authentication Storage & File Systems Middleware Kernel Development Virtualization
Joint Space Operations Center Mission System Application Development Environment
Joint Space Operations Center Mission System Application Development Environment Major Rick Luce 1, Captain Paula Reele 1, ChrisSabol 2, Paul Zetocha 3, Juan Echeverry 4, Richard Kim 4, Barbara Golf 5
Kriterien für ein PetaFlop System
Kriterien für ein PetaFlop System Rainer Keller, HLRS :: :: :: Context: Organizational HLRS is one of the three national supercomputing centers in Germany. The national supercomputing centers are working
Grid Scheduling Dictionary of Terms and Keywords
Grid Scheduling Dictionary Working Group M. Roehrig, Sandia National Laboratories W. Ziegler, Fraunhofer-Institute for Algorithms and Scientific Computing Document: Category: Informational June 2002 Status
locuz.com HPC App Portal V2.0 DATASHEET
locuz.com HPC App Portal V2.0 DATASHEET Ganana HPC App Portal makes it easier for users to run HPC applications without programming and for administrators to better manage their clusters. The web-based
HPC Cluster Decisions and ANSYS Configuration Best Practices. Diana Collier Lead Systems Support Specialist Houston UGM May 2014
HPC Cluster Decisions and ANSYS Configuration Best Practices Diana Collier Lead Systems Support Specialist Houston UGM May 2014 1 Agenda Introduction Lead Systems Support Specialist Cluster Decisions Job
APPENDIX 1 SUBSCRIPTION SERVICES
APPENDIX 1 SUBSCRIPTION SERVICES Red Hat sells subscriptions that entitle you to receive Red Hat services and/or Software during the period of the subscription (generally, one or three years). This Appendix
Recent Advances in HPC for Structural Mechanics Simulations
Recent Advances in HPC for Structural Mechanics Simulations 1 Trends in Engineering Driving Demand for HPC Increase product performance and integrity in less time Consider more design variants Find the
NVIDIA CUDA Software and GPU Parallel Computing Architecture. David B. Kirk, Chief Scientist
NVIDIA CUDA Software and GPU Parallel Computing Architecture David B. Kirk, Chief Scientist Outline Applications of GPU Computing CUDA Programming Model Overview Programming in CUDA The Basics How to Get
Cloud Computing and Enterprise Services
Defense Information Systems Agency A Combat Support Agency Cloud Computing and Enterprise Services Alfred Rivera Technical Program Director 29 July 2010 Peak of Inflated Expectations You Are Here Plateau
Dell Microsoft Business Intelligence and Data Warehousing Reference Configuration Performance Results Phase III
White Paper Dell Microsoft Business Intelligence and Data Warehousing Reference Configuration Performance Results Phase III Performance of Microsoft SQL Server 2008 BI and D/W Solutions on Dell PowerEdge
Cisco Unified Workforce Optimization for Cisco Unified Contact Center Express 8.5
Data Sheet Cisco Unified Workforce Optimization for Cisco Unified Contact Center Express 8.5 Cisco Unified Communications Solutions unify voice, video, data, and mobile applications on fixed and mobile
Networking Modernize. Open Your Network to Innovation
Networking Modernize. Open Your Network to Innovation In a world where real-time information is critical, there s just no room for unnecessary complexity. Did you know? Dell Networking Active Fabric solutions
APPLICATION DEVELOPMENT FOR THE IOT ERA. Embedded Application Development Moves to the Cloud
APPLICATION DEVELOPMENT FOR THE IOT ERA Embedded Application Development Moves to the Cloud EXECUTIVE SUMMARY The emergence of the Internet of Things (IoT) is opening up the complex world of embedded systems
Cluster Scalability of ANSYS FLUENT 12 for a Large Aerodynamics Case on the Darwin Supercomputer
Cluster Scalability of ANSYS FLUENT 12 for a Large Aerodynamics Case on the Darwin Supercomputer Stan Posey, MSc and Bill Loewe, PhD Panasas Inc., Fremont, CA, USA Paul Calleja, PhD University of Cambridge,
Make the Most of Big Data to Drive Innovation Through Reseach
White Paper Make the Most of Big Data to Drive Innovation Through Reseach Bob Burwell, NetApp November 2012 WP-7172 Abstract Monumental data growth is a fact of life in research universities. The ability
Infrastructure as a Service (IaaS)
Infrastructure as a Service (IaaS) DLT Solutions LLC May 2011 Contact Information DLT Cloud Advisory Group 1-855-CLOUD01 (256-8301) [email protected] dl www.dlt.com/cloud Your Hosts Van Ristau Chief Technology
Xeon+FPGA Platform for the Data Center
Xeon+FPGA Platform for the Data Center ISCA/CARL 2015 PK Gupta, Director of Cloud Platform Technology, DCG/CPG Overview Data Center and Workloads Xeon+FPGA Accelerator Platform Applications and Eco-system
SGI High Performance Computing
SGI High Performance Computing Accelerate time to discovery, innovation, and profitability 2014 SGI SGI Company Proprietary 1 Typical Use Cases for SGI HPC Products Large scale-out, distributed memory
icer Bioinformatics Support Fall 2011
icer Bioinformatics Support Fall 2011 John B. Johnston HPC Programmer Institute for Cyber Enabled Research 2011 Michigan State University Board of Trustees. Institute for Cyber Enabled Research (icer)
APPENDIX 1 SUBSCRIPTION SERVICES
APPENDIX 1 SUBSCRIPTION SERVICES Red Hat sells subscriptions that entitle you to receive Red Hat services and/or Software during the period of the subscription (generally, one or three years). This Appendix
Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software
WHITEPAPER Accelerating Enterprise Applications and Reducing TCO with SanDisk ZetaScale Software SanDisk ZetaScale software unlocks the full benefits of flash for In-Memory Compute and NoSQL applications
Journée Mésochallenges 2015 SysFera and ROMEO Make Large-Scale CFD Simulations Only 3 Clicks Away
SysFera and ROMEO Make Large-Scale CFD Simulations Only 3 Clicks Away Benjamin Depardon SysFera Sydney Tekam Tech-Am ING Arnaud Renard ROMEO Manufacturing with HPC 98% of products will be developed digitally
Clusters: Mainstream Technology for CAE
Clusters: Mainstream Technology for CAE Alanna Dwyer HPC Division, HP Linux and Clusters Sparked a Revolution in High Performance Computing! Supercomputing performance now affordable and accessible Linux
LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance
11 th International LS-DYNA Users Conference Session # LS-DYNA Best-Practices: Networking, MPI and Parallel File System Effect on LS-DYNA Performance Gilad Shainer 1, Tong Liu 2, Jeff Layton 3, Onur Celebioglu
Product Framework. Products must use a common suite of infrastructure Support Services
DISA Platform as a Service (PaaS) And RACE May 2012 Version 1.0 Product Framework A Combat Support Agency Overview Enterprise Services must be Product focused with: Discrete retail based boundaries for
How To Improve The Defense Communications System
Briefing to the AFCEA International Cyber Symposium 26 June 2013 Lt Gen Ronnie D. Hawkins, Jr. Director, Defense Information Systems Agency 1 1 UNCLASSIFIED Organizational Changes Command Staff Financial
Broadening Moab/TORQUE for Expanding User Needs
Broadening Moab/TORQUE for Expanding User Needs Gary D. Brown HPC Product Manager CUG 2016 1 2016 Adaptive Computing Enterprises, Inc. Agenda DataWarp Intel MIC KNL Viewpoint Web Portal User Portal Administrator
THE EXPAND PARALLEL FILE SYSTEM A FILE SYSTEM FOR CLUSTER AND GRID COMPUTING. José Daniel García Sánchez ARCOS Group University Carlos III of Madrid
THE EXPAND PARALLEL FILE SYSTEM A FILE SYSTEM FOR CLUSTER AND GRID COMPUTING José Daniel García Sánchez ARCOS Group University Carlos III of Madrid Contents 2 The ARCOS Group. Expand motivation. Expand
Petascale Software Challenges. Piyush Chaudhary [email protected] High Performance Computing
Petascale Software Challenges Piyush Chaudhary [email protected] High Performance Computing Fundamental Observations Applications are struggling to realize growth in sustained performance at scale Reasons
Remote & Collaborative Visualization. Texas Advanced Compu1ng Center
Remote & Collaborative Visualization Texas Advanced Compu1ng Center So6ware Requirements SSH client VNC client Recommended: TigerVNC http://sourceforge.net/projects/tigervnc/files/ Web browser with Java
Parallel Computing with MATLAB
Parallel Computing with MATLAB Scott Benway Senior Account Manager Jiro Doke, Ph.D. Senior Application Engineer 2013 The MathWorks, Inc. 1 Acceleration Strategies Applied in MATLAB Approach Options Best
Cloud Computing with Red Hat Solutions. Sivaram Shunmugam Red Hat Asia Pacific Pte Ltd. [email protected]
Cloud Computing with Red Hat Solutions Sivaram Shunmugam Red Hat Asia Pacific Pte Ltd [email protected] Linux Automation Details Red Hat's Linux Automation strategy for next-generation IT infrastructure
Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes. Anthony Kenisky, VP of North America Sales
Appro Supercomputer Solutions Best Practices Appro 2012 Deployment Successes Anthony Kenisky, VP of North America Sales About Appro Over 20 Years of Experience 1991 2000 OEM Server Manufacturer 2001-2007
HPC and Big Data. EPCC The University of Edinburgh. Adrian Jackson Technical Architect [email protected]
HPC and Big Data EPCC The University of Edinburgh Adrian Jackson Technical Architect [email protected] EPCC Facilities Technology Transfer European Projects HPC Research Visitor Programmes Training
VALUE PROPOSITION FOR SERVICE PROVIDERS. Helping Service Providers accelerate adoption of the cloud
VALUE PROPOSITION FOR SERVICE PROVIDERS Helping Service Providers accelerate adoption of the cloud Partnership with Service Providers Enabling Your Cloud Services in Complex Environments Today s challenge
FLOW-3D Performance Benchmark and Profiling. September 2012
FLOW-3D Performance Benchmark and Profiling September 2012 Note The following research was performed under the HPC Advisory Council activities Participating vendors: FLOW-3D, Dell, Intel, Mellanox Compute
Running Oracle s PeopleSoft Human Capital Management on Oracle SuperCluster T5-8 O R A C L E W H I T E P A P E R L A S T U P D A T E D J U N E 2 0 15
Running Oracle s PeopleSoft Human Capital Management on Oracle SuperCluster T5-8 O R A C L E W H I T E P A P E R L A S T U P D A T E D J U N E 2 0 15 Table of Contents Fully Integrated Hardware and Software
LS DYNA Performance Benchmarks and Profiling. January 2009
LS DYNA Performance Benchmarks and Profiling January 2009 Note The following research was performed under the HPC Advisory Council activities AMD, Dell, Mellanox HPC Advisory Council Cluster Center The
Introduction to ACENET Accelerating Discovery with Computational Research May, 2015
Introduction to ACENET Accelerating Discovery with Computational Research May, 2015 What is ACENET? What is ACENET? Shared regional resource for... high-performance computing (HPC) remote collaboration
CS 698: Special Topics in Big Data. Chapter 2. Computing Trends for Big Data
CS 698: Special Topics in Big Data Chapter 2. Computing Trends for Big Data Chase Wu Associate Professor Department of Computer Science New Jersey Institute of Technology [email protected] Collaborative
SPAWAR HQ ARCHITECTURE AND HUMAN SYSTEMS GROUP Human-Systems Integration/Systems Engineering Support Performance Work Statement
SPAWAR HQ ARCHITECTURE AND HUMAN SYSTEMS GROUP Human-Systems Integration/Systems Engineering Support Performance Work Statement 1.0 INTRODUCTION The Department of the Navy Space and Naval Warfare Systems
AFCEA Aberdeen Luncheon. Army Common Operating Environment (COE) Update. March 11, 2015
AFCEA Aberdeen Luncheon Army Common Operating Environment (COE) Update Mr. Phillip Minor, Deputy Director, COE Directorate Assistant Secretary of the Army for Acquisition, Logistics and Technology (ASA(ALT))
Mitglied der Helmholtz-Gemeinschaft. System monitoring with LLview and the Parallel Tools Platform
Mitglied der Helmholtz-Gemeinschaft System monitoring with LLview and the Parallel Tools Platform November 25, 2014 Carsten Karbach Content 1 LLview 2 Parallel Tools Platform (PTP) 3 Latest features 4
Scaling from Workstation to Cluster for Compute-Intensive Applications
Cluster Transition Guide: Scaling from Workstation to Cluster for Compute-Intensive Applications IN THIS GUIDE: The Why: Proven Performance Gains On Cluster Vs. Workstation The What: Recommended Reference
Cisco Unified Workforce Optimization for Cisco Unified Contact Center Express 8.0
Cisco Unified Workforce Optimization for Express 8.0 Cisco Unified Communications Solutions unify voice, video, data, and mobile applications on fixed and mobile networks, enabling easy collaboration every
Private cloud computing advances
Building robust private cloud services infrastructures By Brian Gautreau and Gong Wang Private clouds optimize utilization and management of IT resources to heighten availability. Microsoft Private Cloud
How To Develop A Data Center Cost Model
Considerations in Data Center Migration Life Cycle Cost Estimating A Joint Information Environment (JIE) Perspective Dennis J. McKeon June 2015 Approved for Public Release; Distribution Unlimited 15-1446.
Overview of Requirements and Applications for 40 Gigabit and 100 Gigabit Ethernet
Overview of Requirements and Applications for 40 Gigabit and 100 Gigabit Ethernet Version 1.1 June 2010 Authors: Mark Nowell, Cisco Vijay Vusirikala, Infinera Robert Hays, Intel 1. This work represents
Cluster, Grid, Cloud Concepts
Cluster, Grid, Cloud Concepts Kalaiselvan.K Contents Section 1: Cluster Section 2: Grid Section 3: Cloud Cluster An Overview Need for a Cluster Cluster categorizations A computer cluster is a group of
Understanding the Benefits of IBM SPSS Statistics Server
IBM SPSS Statistics Server Understanding the Benefits of IBM SPSS Statistics Server Contents: 1 Introduction 2 Performance 101: Understanding the drivers of better performance 3 Why performance is faster
Title. Click to edit Master text styles Second level Third level
Title Click to edit Master text styles Second level Third level IBM s Vision For The New Enterprise Data Center Subram Natarajan Senior Consultant, STG Asia Pacific [email protected] Multiple
Deploying and managing a Visualization Farm @ Onera
Deploying and managing a Visualization Farm @ Onera Onera Scientific Day - October, 3 2012 Network and computing department (DRI), Onera P.F. Berte [email protected] Plan Onera global HPC
