Getting OpenMP Up To Speed



Similar documents
Parallel Computing. Shared memory parallel programming with OpenMP

MPI and Hybrid Programming Models. William Gropp

OpenMP and Performance

Parallel Computing. Parallel shared memory computing with OpenMP

OpenMP & MPI CISC 879. Tristan Vanderbruggen & John Cavazos Dept of Computer & Information Sciences University of Delaware

#pragma omp critical x = x + 1; !$OMP CRITICAL X = X + 1!$OMP END CRITICAL. (Very inefficiant) example using critical instead of reduction:

Parallel Algorithm Engineering

<Insert Picture Here> An Experimental Model to Analyze OpenMP Applications for System Utilization

Spring 2011 Prof. Hyesoon Kim

COMP/CS 605: Introduction to Parallel Computing Lecture 21: Shared Memory Programming with OpenMP

Parallel Programming Survey

Multi-Threading Performance on Commodity Multi-Core Processors

Comparing the OpenMP, MPI, and Hybrid Programming Paradigm on an SMP Cluster

OpenMP Programming on ScaleMP

OpenACC 2.0 and the PGI Accelerator Compilers

An Introduction to Parallel Programming with OpenMP

A Comparison Of Shared Memory Parallel Programming Models. Jace A Mogill David Haglin

Introduction. What is RAID? The Array and RAID Controller Concept. Click here to print this article. Re-Printed From SLCentral

High Performance Computing. Course Notes HPC Fundamentals

Introduction to OpenMP Programming. NERSC Staff

High Performance Computing

Performance Evaluation of NAS Parallel Benchmarks on Intel Xeon Phi

Q & A From Hitachi Data Systems WebTech Presentation:

Assessing the Performance of OpenMP Programs on the Intel Xeon Phi

What is Multi Core Architecture?

22S:295 Seminar in Applied Statistics High Performance Computing in Statistics

Lecture 4. Parallel Programming II. Homework & Reading. Page 1. Projects handout On Friday Form teams, groups of two

Scalability evaluation of barrier algorithms for OpenMP

Quiz for Chapter 1 Computer Abstractions and Technology 3.10

MAQAO Performance Analysis and Optimization Tool

To copy all examples and exercises to your local scratch directory type: /g/public/training/openmp/setup.sh

Scheduling Task Parallelism" on Multi-Socket Multicore Systems"

THE NAS KERNEL BENCHMARK PROGRAM

Workshare Process of Thread Programming and MPI Model on Multicore Architecture

Parallel Computing using MATLAB Distributed Compute Server ZORRO HPC

Chapter 7: Distributed Systems: Warehouse-Scale Computing. Fall 2011 Jussi Kangasharju

Debugging with TotalView

BLM 413E - Parallel Programming Lecture 3

Hybrid Programming with MPI and OpenMP

Evaluation of CUDA Fortran for the CFD code Strukti

Contributions to Gang Scheduling

Multicore Parallel Computing with OpenMP

OpenACC Programming and Best Practices Guide

Multi-core architectures. Jernej Barbic , Spring 2007 May 3, 2007

Lecture 3: Evaluating Computer Architectures. Software & Hardware: The Virtuous Cycle?

Performance Analysis and Optimization Tool

Programming the Intel Xeon Phi Coprocessor

reduction critical_section

OpenMP Application Program Interface

Garbage Collection in the Java HotSpot Virtual Machine

A Case Study - Scaling Legacy Code on Next Generation Platforms

LS-DYNA Scalability on Cray Supercomputers. Tin-Ting Zhu, Cray Inc. Jason Wang, Livermore Software Technology Corp.

David Rioja Redondo Telecommunication Engineer Englobe Technologies and Systems

OpenMP* 4.0 for HPC in a Nutshell

Objectives. Overview of OpenMP. Structured blocks. Variable scope, work-sharing. Scheduling, synchronization

Why Computers Are Getting Slower (and what we can do about it) Rik van Riel Sr. Software Engineer, Red Hat

Using the Intel Inspector XE

Programming Languages for Large Scale Parallel Computing. Marc Snir

Whitepaper: performance of SqlBulkCopy

2: Computer Performance

A Pattern-Based Comparison of OpenACC & OpenMP for Accelerators

Parallelism and Cloud Computing

Overview on Modern Accelerators and Programming Paradigms Ivan Giro7o

Virtuoso and Database Scalability

Real-Time Systems Prof. Dr. Rajib Mall Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur

How To Understand The Concept Of A Distributed System

OpenMP C and C++ Application Program Interface

Accelerating sequential computer vision algorithms using OpenMP and OpenCL on commodity parallel hardware

UTS: An Unbalanced Tree Search Benchmark

Load Imbalance Analysis

RevoScaleR Speed and Scalability

A Simultaneous Solution for General Linear Equations on a Ring or Hierarchical Cluster

Overlapping Data Transfer With Application Execution on Clusters

Parallelization of video compressing with FFmpeg and OpenMP in supercomputing environment

PROACTIVE BOTTLENECK PERFORMANCE ANALYSIS IN PARALLEL COMPUTING USING OPENMP

Real-Time Scheduling 1 / 39

5x in 5 hours Porting SEISMIC_CPML using the PGI Accelerator Model

ETHEL THE AARDVARK GOES BGP ROUTING

Understanding Hardware Transactional Memory

Keys to node-level performance analysis and threading in HPC applications

Experiences with HPC on Windows

Chapter 2 Parallel Architecture, Software And Performance

Architecture of Hitachi SR-8000

GPU File System Encryption Kartik Kulkarni and Eugene Linkov

An Implementation Of Multiprocessor Linux

Transcription:

1 Getting OpenMP Up To Speed Ruud van der Pas Senior Staff Engineer Oracle Solaris Studio Oracle Menlo Park, CA, USA IWOMP 2010 CCS, University of Tsukuba Tsukuba, Japan June 14-16, 2010

2 Outline The Myth Deep Trouble Get Real The Wrapping

3 A myth, in popular use, is something that is widely believed but false.... (source: www.reference.com)

4 The Myth OpenMP Does Not Scale

5 Hmmm... What Does That Really Mean?

6 Some Questions I Could Ask Do you mean you wrote a parallel program, using OpenMP and it doesn't perform? I see. Did you make sure the program was fairly well optimized in sequential mode? Oh. You didn't. By the way, why do you expect the program to scale? Oh. You just think it should and you used all the cores. Have you estimated the speed up using Amdahl's Law? No, this law is not a new EU environmental regulation. It is something else. I understand. You can't know everything. Have you at least used a tool to identify the most time consuming parts in your program?

7 Some More Questions I Could Ask Oh. You didn't. You just parallelized all loops in the program. Did you try to avoid parallelizing innermost loops in a loop nest? Oh. You didn't. Did you minimize the number of parallel regions then? Oh. You didn't. It just worked fine the way it was. Did you at least use the nowait clause to minimize the use of barriers? Oh. You've never heard of a barrier. Might be worth to read up on. Do all processors roughly perform the same amount of work? You don't know, but think it is okay. I hope you're right.

8 I Don't Give Up That Easily Did you make optimal use of private data, or did you share most of it? Oh. You didn't. Sharing is just easier. I see. You seem to be using a cc-numa system. Did you take that into account? You've never heard of that either. How unfortunate. Could there perhaps be any false sharing affecting performance? Oh. Never heard of that either. May come handy to learn a little more about both. So, what did you do next to address the performance? Switched to MPI. Does that perform any better then? Oh. You don't know. You're still debugging the code.

9 Going Into Pedantic Mode While you're waiting for your MPI debug run to finish (are you sure it doesn't hang by the way), please allow me to talk a little more about OpenMP and Performance.

10 Deep Trouble

11 OpenMP and Performance The transparency of OpenMP is a mixed blessing Makes things pretty easy May mask performance bottlenecks In the ideal world, an OpenMP application just performs well Unfortunately, this is not the case Two of the more obscure effects that can negatively impact performance are cc-numa behavior and False Sharing Neither of these are restricted to OpenMP, but they are important enough to cover in some detail here

12 False Sharing

13 False Sharing A store into a shared cache line invalidates the other copies of that line: CPUs Caches Memory The system is not able to distinguish between changes within one individual line

14 False Sharing Red Flags Be alert, when all of these three conditions are met: Shared data is modifi ed by multiple processors Multiple threads operate on the same cache line(s) Update occurs simultaneously and very frequently Use local data where possible Shared read-only data does not lead to false sharing

15 Considerations for cc-numa

Memory 16 A generic cc-numa architecture Memory Processor Processor Local Access (fast) Cache Coherent Interconnect Remote Access (slower) Main Issue: How To Distribute The Data?

17 About Data Distribution Important aspect on a cc-numa system If not optimal - longer access times, memory hotspots OpenMP does not provide support for cc-numa Placement comes from the Operating System This is therefore Operating System dependent Solaris, Linux and Windows use First Touch to place data

Memory 18 About First Touch placement/1 a[0] : a[99] Memory Processor Processor Cache Coherent Interconnect for (i=0; i<100; i++) a[i] = 0; First Touch All array elements are in the memory of the processor executing this thread

Memory 19 About First Touch placement/2 a[0] : a[49] Memory Processor Processor a[50] : a[99] Cache Coherent Interconnect #pragma omp parallel for num_threads(2) for (i=0; i<100; i++) a[i] = 0; First Touch Both memories each have their half of the array

20 Get Real

21 Block Matrix Update

22 A 3D matrix update do k = 2, n do j = 2, n!$omp parallel do default(shared) private(i) &!$omp schedule(static) do i = 1, m x(i,j,k) = x(i,j,k-1) + x(i,j-1,k)*scale end do!$omp end parallel do end do end do The loops are correctly nested for serial performance Due to a data dependency on J and K, only the inner loop can be parallelized This will cause the barrier to be executed (N-1) 2 times I K k k-1 Data Dependency Graph j-1 j J

23 The performance Performance (Mfl op/s) 350 300 250 200 150 100 50 Inner loop over I has been parallelized Scaling is very poor (as to be expected) 0 0 10 20 30 40 50 60 Number of threads Dimensions : M=7,500 N=20 Footprint : ~24 MByte

scales somewhat 24 Performance Analyzer data Name Excl. User Incl. Excl. Using 10 threads CPU User CPU Wall sec. % sec. sec. <Total> 10.590 100.0 10.590 1.550 mt_endoftask_barrier_ 5.740 54.2 5.740 0.240 mt_waitforwork_ 3.860 36.4 3.860 0. mt_masterfunction_ 0.480 4.5 0.680 0.480 MAIN_ 0.230 2.2 1.200 0.470 block_3d_ -- MP doall from line 14 [_$d1a14.block_3d_] 0.170 1.6 5.910 0.170 block_3d_ 0.040 0.4 6.460 0.040 memset 0.030 0.3 0.030 0.080 Using 20 threads do not scale at all Name Excl. User Incl. Excl. CPU User CPU Wall sec. % sec. sec. <Total> 47.120 100.0 47.120 2.900 mt_endoftask_barrier_ 25.700 54.5 25.700 0.980 mt_waitforwork_ 19.880 42.2 19.880 0. mt_masterfunction_ 1.100 2.3 1.320 1.100 MAIN_ 0.190 0.4 2.520 0.470 block_3d_ -- MP doall from line 14 [_$d1a14.block_3d_] 0.100 0.2 25.800 0.100 mt_setup_dojob_int_ 0.080 0.2 0.080 0.080 mt_setup_job_ 0.020 0.0 0.020 0.020 block_3d_ 0.010 0.0 27.020 0.010 Question: Why is mt_waitforwork so high in the profi le?

25 The Analyzer Timeline overview mt_waitforwork mt_endoftask_barrier

26 This is False Sharing at work!!$omp parallel do default(shared) private(i) &!$omp schedule(static) do i = 1, m x(i,j,k) = x(i,j,k-1) + x(i,j-1,k)*scale end do!$omp end parallel do P=1 P=2 P=4 P=8 no sharing False sharing increases as we increase the number of threads

27 Sanity Check: Setting M=75000* Only a very few barrier calls now *) Increasing the length of the loop should decrease false sharing

28 Performance comparison 700 Performance (Mfl op/s) 600 500 400 300 200 100 M = 7,500 M = 75,000 0 0 10 20 30 40 50 60 Number of threads For a higher value of M, the program scales better

29 Observation K No data dependency on 'I' Therefore we can split the 3D matrix in larger blocks and process these in parallel I J do k = 2, n do j = 2, n do i = 1, m x(i,j,k) = x(i,j,k-1) + x(i,j-1,k)*scale end do end do end do

30 The Idea I ie is K We need to distribute the M iterations over the number of processors We do this by controlling the start (IS) and end (IE) value of the inner loop J Each thread will calculate these values for it's portion of the work do k = 2, n do j = 2, n do i = is, ie x(i,j,k) = x(i,j,k-1) + x(i,j-1,k)*scale end do end do end do

31 The fi rst implementation use omp_lib... nrem = mod(m,nthreads) nchunk = (m-nrem)/nthreads!$omp parallel default (none)&!$omp private (P,is,ie) &!$omp shared (nrem,nchunk,m,n,x,scale) subroutine kernel(is,ie,m,n,x,scale)... do k = 2, n do j = 2, n do i = is, ie x(i,j,k)=x(i,j,k-1)+x(i,j-1,k)*scale end do end do end do P = omp_get_thread_num() if ( P < nrem ) then is = 1 + P*(nchunk + 1) ie = is + nchunk else is = 1 + P*nchunk+ nrem ie = is + nchunk - 1 end if call kernel(is,ie,m,n,x,scale)!$omp end parallel

32 Another Idea: Use OpenMP! use omp_lib implicit none integer :: is, ie, m, n real(kind=8):: x(m,n,n), scale integer :: i, j, k!$omp parallel default(none) &!$omp private(i,j,k) shared(m,n,scale,x) do k = 2, n do j = 2, n!$omp do schedule(static) do i = 1, m x(i,j,k) = x(i,j,k-1) + x(i,j-1,k)*scale end do!$omp end do nowait end do end do!$omp end parallel

33 How this works on 2 threads Thread 0 Executes: Thread 1 Executes: k=2 j=2 parallel region k=2 j=2 do i = 1,m/2 x(i,2,2) =... end do work sharing do i = m/2+1,m x(i,2,2) =... end do k=2 j=3 parallel region k=2 j=3 do i = 1,m/2 x(i,3,2) =... end do... etc... work sharing do i = m/2+1,m x(i,3,2) =... end do This splits the operation in a way... etc that... is similar to our manual implementation

34 Performance We have set M=7500 N=20 This problem size does not scale at all when we explicitly parallelized the inner loop over 'I' We have have tested 4 versions of this program Inner Loop Over 'I' - Our fi rst OpenMP version AutoPar - The automatically parallelized version of 'kernel' OMP_Chunks - The manually parallelized version with our explicit calculation of the chunks OMP_DO - The version with the OpenMP parallel region and work-sharing DO

35 The performance (M=7,500) Performance (Mfl op/s) 2500 2000 1500 1000 500 OMP DO OMP Chunks Innerloop The auto-parallelizing compiler does really well! 0 0 10 20 30 40 50 60 Number of threads Dimensions : M=7,500 N=20 Footprint : ~24 MByte

36 Matrix Times Vector

37 The Sequential Source for (i=0; i<m; i++) { a[i] = 0.0; for (j=0; j<n; j++) a[i] += b[i][j]*c[j]; } i = * j

38 The OpenMP Source #pragma omp parallel for default(none) \ private(i,j) shared(m,n,a,b,c) j for (i=0; i<m; i++) { a[i] = 0.0; = * for (j=0; j<n; j++) a[i] += b[i][j]*c[j]; i }

39 Performance - 2 Socket Nehalem 35000 1 Thread 30000 2 Threads 4 Threads Performance (Mflop/s) 25000 20000 15000 10000 8 Threads 16 Threads Wait a minute, this operation is highly parallel... Speed-up is ~1.6x only 5000 0 0.1 1 10 100 1000 10000 100000 1000000 Memory Footprint (KByte)

socket 0 socket 1 40 A Two Socket Nehalem System Processor Number QPI Interconnect memory memory shared cache shared cache caches caches caches caches caches caches caches caches core 0 core 1 core 2 core 3 core 0 core 1 core 2 core 3 hw thread 0 hw thread 1 hw thread 0 hw thread 1 hw thread 0 hw thread 1 hw thread 0 hw thread 1 hw thread 0 hw thread 1 hw thread 0 hw thread 1 hw thread 0 hw thread 1 hw thread 0 hw thread 1 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15

41 Data Initialization #pragma omp parallel default(none) \ shared(m,n,a,b,c) private(i,j) { #pragma omp for for (j=0; j<n; j++) c[j] = 1.0; j #pragma omp for for (i=0; i<m; i++) { a[i] = -1957.0; for (j=0; j<n; j++) b[i[]j] = i; } /*-- End of omp for --*/ i = * } /*-- End of parallel region --*/

42 Exploit First Touch 35000 30000 1 Thread 2 Threads Performance (Mflop/s) 25000 20000 15000 10000 4 Threads 8 Threads 16 Threads The only change is the way the data is distributed over the system Max speed up is ~3.2x 5000 0 0.1 1 10 100 1000 10000 100000 1000000 Memory Footprint (KByte)

43 Summary Case Studies There are several important basic aspects to consider when it comes to writing an effi cient OpenMP program Moreover, there are also obscure additional aspects: cc-numa False Sharing Key problem is that most developers are not aware of these rules and... blaming OpenMP is all that easy In some cases it is a trade-off between ease of use and performance OpenMP typically goes for the former, but... With some extra effort can be made to scale well in many cases

44 The Wrapping

45 Wrapping Things Up While we're still waiting for your MPI debug run to finish, I want to ask you whether you found my information useful. Yes, it is overwhelming. I know. And OpenMP is somewhat obscure in certain areas. I know that as well. I understand. You're not a Computer Scientist and just need to get your scientific research done. I agree this is not a good situation, but it is all about Darwin, you know. I'm sorry, it is a tough world out there.

46 It Never Ends Oh, your MPI job just finished! Great. Your program does not write a file called 'core' and it wasn't there when you started the program? You wonder where such a file comes from? Let's get a big and strong coffee first.

47 That's It Thank You and... Stay Tuned! Ruud van der Pas ruud.vanderpas@sun.com