An Introduction to Modern GPU Architecture Ashu Rege Director of Developer Technology

Similar documents
The Evolution of Computer Graphics. SVP, Content & Technology, NVIDIA

GPGPU Computing. Yong Cao

Radeon HD 2900 and Geometry Generation. Michael Doggett

Computer Graphics Hardware An Overview

GPU(Graphics Processing Unit) with a Focus on Nvidia GeForce 6 Series. By: Binesh Tuladhar Clay Smith

GPU Architecture. Michael Doggett ATI

NVIDIA GeForce GTX 580 GPU Datasheet

Introduction to GPGPU. Tiziano Diamanti

Introduction to GP-GPUs. Advanced Computer Architectures, Cristina Silvano, Politecnico di Milano 1

Real-Time Realistic Rendering. Michael Doggett Docent Department of Computer Science Lund university

Shader Model 3.0. Ashu Rege. NVIDIA Developer Technology Group

Recent Advances and Future Trends in Graphics Hardware. Michael Doggett Architect November 23, 2005

Graphics Cards and Graphics Processing Units. Ben Johnstone Russ Martin November 15, 2011

Next Generation GPU Architecture Code-named Fermi

Lecture 11: Multi-Core and GPU. Multithreading. Integration of multiple processor cores on a single chip.

OpenCL Optimization. San Jose 10/2/2009 Peng Wang, NVIDIA

Radeon GPU Architecture and the Radeon 4800 series. Michael Doggett Graphics Architecture Group June 27, 2008

Introduction to GPU Architecture

Lecture 3: Modern GPUs A Hardware Perspective Mohamed Zahran (aka Z) mzahran@cs.nyu.edu

OpenGL Performance Tuning

L20: GPU Architecture and Models

This Unit: Putting It All Together. CIS 501 Computer Architecture. Sources. What is Computer Architecture?

GPU Parallel Computing Architecture and CUDA Programming Model

Introduction to GPU Programming Languages

ATI Radeon 4800 series Graphics. Michael Doggett Graphics Architecture Group Graphics Product Group

GPUs Under the Hood. Prof. Aaron Lanterman School of Electrical and Computer Engineering Georgia Institute of Technology

GPU Architectures. A CPU Perspective. Data Parallelism: What is it, and how to exploit it? Workload characteristics

Introduction GPU Hardware GPU Computing Today GPU Computing Example Outlook Summary. GPU Computing. Numerical Simulation - from Models to Software

GPU Hardware and Programming Models. Jeremy Appleyard, September 2015

Image Synthesis. Transparency. computer graphics & visualization

GPU System Architecture. Alan Gray EPCC The University of Edinburgh

Image Processing and Computer Graphics. Rendering Pipeline. Matthias Teschner. Computer Science Department University of Freiburg

Touchstone -A Fresh Approach to Multimedia for the PC

Dynamic Resolution Rendering

Overview. Lecture 1: an introduction to CUDA. Hardware view. Hardware view. hardware view software view CUDA programming

GPU Computing with CUDA Lecture 2 - CUDA Memories. Christopher Cooper Boston University August, 2011 UTFSM, Valparaíso, Chile

Introduction to GPU hardware and to CUDA

3D Computer Games History and Technology

Advanced Visual Effects with Direct3D

AMD GPU Architecture. OpenCL Tutorial, PPAM Dominik Behr September 13th, 2009

CUDA programming on NVIDIA GPUs

GPUs for Scientific Computing

INTRODUCTION TO RENDERING TECHNIQUES

GPU Architecture. An OpenCL Programmer s Introduction. Lee Howes November 3, 2010

Optimizing Unity Games for Mobile Platforms. Angelo Theodorou Software Engineer Unite 2013, 28 th -30 th August

Midgard GPU Architecture. October 2014

Optimizing AAA Games for Mobile Platforms

GPU Computing - CUDA

A Crash Course on Programmable Graphics Hardware

NVIDIA GeForce GTX 750 Ti

How To Use An Amd Ramfire R7 With A 4Gb Memory Card With A 2Gb Memory Chip With A 3D Graphics Card With An 8Gb Card With 2Gb Graphics Card (With 2D) And A 2D Video Card With

NVIDIA workstation 3D graphics card upgrade options deliver productivity improvements and superior image quality

Applications to Computational Financial and GPU Computing. May 16th. Dr. Daniel Egloff

GPGPU: General-Purpose Computation on GPUs

QCD as a Video Game?

Introduction to Computer Graphics

OpenCL Programming for the CUDA Architecture. Version 2.3

Parallel Programming Survey

GPU architecture II: Scheduling the graphics pipeline

SAPPHIRE HD GB GDDR5 PCIE.

CSE 564: Visualization. GPU Programming (First Steps) GPU Generations. Klaus Mueller. Computer Science Department Stony Brook University

CUDA Optimization with NVIDIA Tools. Julien Demouth, NVIDIA

Introduction to Cloud Computing

LBM BASED FLOW SIMULATION USING GPU COMPUTING PROCESSOR

NVIDIA CUDA Software and GPU Parallel Computing Architecture. David B. Kirk, Chief Scientist

The Future Of Animation Is Games

Awards News. GDDR5 memory provides twice the bandwidth per pin of GDDR3 memory, delivering more speed and higher bandwidth.

General Purpose Computation on Graphics Processors (GPGPU) Mike Houston, Stanford University

ultra fast SOM using CUDA

Choosing a Computer for Running SLX, P3D, and P5

Graphics and Computing GPUs

1. INTRODUCTION Graphics 2

White Paper AMD GRAPHICS CORES NEXT (GCN) ARCHITECTURE

Hardware design for ray tracing

Maximize Application Performance On the Go and In the Cloud with OpenCL* on Intel Architecture

Texture Cache Approximation on GPUs

Data Parallel Computing on Graphics Hardware. Ian Buck Stanford University

E6895 Advanced Big Data Analytics Lecture 14:! NVIDIA GPU Examples and GPU on ios devices

FLOATING-POINT ARITHMETIC IN AMD PROCESSORS MICHAEL SCHULTE AMD RESEARCH JUNE 2015

Real-Time Graphics Architecture

2: Introducing image synthesis. Some orientation how did we get here? Graphics system architecture Overview of OpenGL / GLU / GLUT

SAPPHIRE TOXIC R9 270X 2GB GDDR5 WITH BOOST

Stream Processing on GPUs Using Distributed Multimedia Middleware


CUDA. Multicore machines

Architecture of Hitachi SR-8000

Optimization for DirectX9 Graphics. Ashu Rege

DATA VISUALIZATION OF THE GRAPHICS PIPELINE: TRACKING STATE WITH THE STATEVIEWER

Introduction to Game Programming. Steven Osman

Guided Performance Analysis with the NVIDIA Visual Profiler

GPGPU for Real-Time Data Analytics: Introduction. Nanyang Technological University, Singapore 2

Accelerating Intensity Layer Based Pencil Filter Algorithm using CUDA

Advanced Rendering for Engineering & Styling

Transcription:

An Introduction to Modern GPU Architecture Ashu Rege Director of Developer Technology

Agenda Evolution of GPUs Computing Revolution Stream Processing Architecture details of modern GPUs

Evolution of GPUs

Evolution of GPUs (1995-1999) 1995 NV1 1997 Riva 128 (NV3), DX3 1998 Riva TNT (NV4), DX5 32 bit color, 24 bit Z, 8 bit stencil Dual texture, bilinear filtering 2 pixels per clock (ppc) 1999 Riva TNT2 (NV5), DX6 Faster TNT 128b memory interface 32 MB memory The chip that would not die Virtua Fighter (SEGA Corporation) NV1 50K triangles/sec 1M pixel ops/sec 1M transistors 16-bit color Nearest filtering 1995

Evolution of GPUs (Fixed Function) GeForce 256 (NV10) DirectX 7.0 Hardware T&L Cubemaps DOT3 bump mapping Register combiners 2x Anisotropic filtering Trilinear filtering DXT texture compression 4 ppc Term GPU introduced Deus Ex (Eidos/Ion Storm) NV10 15M triangles/sec 480M pixel ops/sec 23M transistors 32-bit color Trilinear filtering 1999

NV10 Register Combiners Input RGB, Alpha Registers Input Mappings RGB Function RGB Scale/Bias Next Combiner s RGB Registers A A op 1 B RGB Portion B C C op 2 D D AB op 3 CD Input Alpha, Blue Registers Input Mappings Alpha Function Alpha Scale/Bias Next Combiner s Alpha Registers A AB Alpha Portion B C CD D AB op 4 CD

Evolution of GPUs (Shader Model 1.0) GeForce 3 (NV20) NV2A Xbox GPU DirectX 8.0 Vertex and Pixel Shaders 3D Textures Hardware Shadow Maps 8x Anisotropic filtering Multisample AA (MSAA) 4 ppc Ragnarok Online (Atari/Gravity) NV20 100M triangles/sec 1G pixel ops/sec 57M transistors Vertex/Pixel shaders MSAA 2001

Evolution of GPUs (Shader Model 2.0) GeForce FX Series (NV3x) DirectX 9.0 Floating Point and Long Vertex and Pixel Shaders Shader Model 2.0 256 vertex ops 32 tex + 64 arith pixel ops Shader Model 2.0a 256 vertex ops Up to 512 ops Shading Languages HLSL, Cg, GLSL Dawn Demo (NVIDIA) NV30 200M triangles/sec 2G pixel ops/sec 125M transistors Shader Model 2.0a 2003

Evolution of GPUs (Shader Model 3.0) GeForce 6 Series (NV4x) DirectX 9.0c Shader Model 3.0 Dynamic Flow Control in Vertex and Pixel Shaders 1 Branching, Looping, Predication, Vertex Texture Fetch High Dynamic Range (HDR) 64 bit render target FP16x4 Texture Filtering and Blending 1 Some flow control first introduced in SM2.0a Far Cry HDR (Ubisoft/Crytek Crytek) NV40 600M triangles/sec 12.8G pixel ops/sec 220M transistors Shader Model 3.0 Rotated Grid MSAA 16x Aniso,, SLI 2004

Far Cry No HDR/HDR Comparison

Evolution of GPUs (Shader Model 4.0) GeForce 8 Series (G8x) DirectX 10.0 Shader Model 4.0 Geometry Shaders No caps bits Unified Shaders New Driver Model in Vista CUDA based GPU computing GPUs become true computing processors measured in GFLOPS Crysis (EA/Crytek Crytek) G80 Unified Shader Cores w/ Stream Processors 681M transistors Shader Model 4.0 8x MSAA, CSAA 2006

Crysis. Images courtesy of Crytek.

As Of Today GeForce GTX 280 (GT200) DX10 1.4 billion transistors 576 mm 2 in 65nm CMOS 240 stream processors 933 GFLOPS peak 1.3GHz processor clock 1GB DRAM 512 pin DRAM interface 142 GB/s peak

Stunning Graphics Realism Lush, Rich Worlds Crysis 2006 Crytek / Electronic Arts Hellgate: London 2005-2006 Flagship Studios, Inc. Licensed by NAMCO BANDAI Games America, Inc. Incredible Physics Effects Core of the Definitive Gaming Platform

What Is Behind This Computing Revolution? Unified Scalar Shader Architecture Highly Data Parallel Stream Processing Next, let s try to understand what these terms mean

Unified Scalar Shader Architecture

Graphics Pipelines For Last 20 Years Processor per function Vertex T&L evolved to vertex shading Triangle Triangle, point, line setup Pixel Flat shading, texturing, eventually pixel shading ROP Blending, Z-buffering, antialiasing Memory Wider and faster over the years

Shaders in Direct3D DirectX 9: Vertex Shader, Pixel Shader DirectX 10: Vertex Shader, Geometry Shader, Pixel Shader DirectX 11: Vertex Shader, Hull Shader, Domain Shader, Geometry Shader, Pixel Shader, Compute Shader Observation: All of these shaders require the same basic functionality: Texturing (or Data Loads) and Math Ops.

Unified Pipeline Physics Geometry (new in DX10) Future Texture + Floating Point Processor ROP Vertex Pixel Compute (CUDA, DX11 Compute, OpenCL) Memory

Why Unify? Vertex Shader Pixel Shader Idle hardware Vertex Shader Idle hardware Unbalanced and inefficient utilization in nonunified architecture Heavy Geometry Workload Perf = 4 Pixel Shader Heavy Pixel Workload Perf = 8

Why Unify? Unified Shader Vertex Workload Pixel Optimal utilization In unified architecture Heavy Geometry Workload Perf = 11 Unified Shader Pixel Workload Vertex Heavy Pixel Workload Perf = 11

Why Scalar Instruction Shader (1) Vector ALU efficiency varies 4 MAD r2.xyzw, r0.xyzw, r1.xyzw 100% utilization 3 DP3 r2.w, r0.xyz, r1.xyz 75% 2 MUL r2.xy, r0.xy, r1.xy 50% 1 ADD r2.w, r0.x, r1.x 25%

Why Scalar Instruction Shader (2) Vector ALU with co-issue better but not perfect DP3 r2.x, r0.xyz, r1.xyz 4 ADD r2.w, r0.w, r1.w 3 DP3 r2.w, r0.xyz, r1.xyz 1 ADD r2.w, r0.w, r2.w } 100% Cannot co-issue Vector/VLIW architecture More compiler work required G8x, GT200: scalar always 100% efficient, simple to compile Up to 2x effective throughput advantage relative to vector

Complex Shader Performance on Scalar Arch. Procedural Perlin Noise Fire 5 Procedural Fire 4.5 4 3.5 3 2.5 2 1.5 1 0.5 0 7900GTX 8800GTX

Conclusion Build a unified architecture with scalar cores where all shader operations are done on the same processors

Stream Processing

The Supercomputing Revolution (1)

The Supercomputing Revolution (2)

What Accounts For This Difference? Need to understand how CPUs and GPUs differ Latency Intolerance versus Latency Tolerance Task Parallelism versus Data Parallelism Multi-threaded Cores versus SIMT (Single Instruction Multiple Thread) Cores 10s of Threads versus 10,000s of Threads

Latency and Throughput Latency is a time delay between the moment something is initiated, and the moment one of its effects begins or becomes detectable For example, the time delay between a request for texture reading and texture data returns Throughput is the amount of work done in a given amount of time For example, how many triangles processed per second CPUs are low latency low throughput processors GPUs are high latency high throughput processors

Latency (1) GPUs are designed for tasks that can tolerate latency Example: Graphics in a game (simplified scenario): CPU Generate Frame 0 Generate Frame 1 Generate Frame 2 GPU Idle Render Frame 0 Render Frame 1 Latency between frame generation and rendering (order of milliseconds) To be efficient, GPUs must have high throughput, i.e. processing millions of pixels in a single frame

Latency (2) CPUs are designed to minimize latency Example: Mouse or keyboard input Caches are needed to minimize latency CPUs are designed to maximize running operations out of cache Instruction pre-fetch Out-of-order execution, flow control CPUs need a large cache, GPUs do not GPUs can dedicate more of the transistor area to computation horsepower

CPU versus GPU Transistor Allocation GPUs can have more ALUs for the same sized chip and therefore run many more threads of computation Control ALU ALU ALU ALU Cache DRAM DRAM CPU GPU Modern GPUs run 10,000s of threads concurrently

Managing Threads On A GPU How do we: Avoid synchronization issues between so many threads? Dispatch, schedule, cache, and context switch 10,000s of threads? Program 10,000s of threads? Design GPUs to run specific types of threads: Independent of each other no synchronization issues SIMD (Single Instruction Multiple Data) threads minimize thread management Reduce hardware overhead for scheduling, caching etc. Program blocks of threads (e.g. one pixel shader per draw call, or group of pixels) Any problems which can be solved with this type of computation?

Data Parallel Problems Plenty of problems fall into this category (luckily ) Graphics, image & video processing, physics, scientific computing, This type of parallelism is called data parallelism And GPUs are the perfect solution for them! In fact the more the data, the more efficient GPUs become at these algorithms Bonus: You can relatively easily add more processing cores to a GPU and increase the throughput

Parallelism in CPUs v. GPUs CPUs use task parallelism Multiple tasks map to multiple threads Tasks run different instructions 10s of relatively heavyweight threads run on 10s of cores Each thread managed and scheduled explicitly Each thread has to be individually programmed GPUs use data parallelism SIMD model (Single Instruction Multiple Data) Same instruction on different data 10,000s of lightweight threads on 100s of cores Threads are managed and scheduled by hardware Programming done for batches of threads (e.g. one pixel shader per group of pixels, or draw call)

Stream Processing What we just described: Given a (typically large) set of data ( stream ) Run the same series of operations ( kernel or shader ) on all of the data (SIMD) GPUs use various optimizations to improve throughput: Some on-chip memory and local caches to reduce bandwidth to external memory Batch groups of threads to minimize incoherent memory access Bad access patterns will lead to higher latency and/or thread stalls. Eliminate unnecessary operations by exiting or killing threads Example: Z-Culling and Early-Z to kill pixels which will not be displayed

To Summarize GPUs use stream processing to achieve high throughput GPUs designed to solve problems that tolerate high latencies High latency tolerance Lower cache requirements Less transistor area for cache More area for computing units More computing units 10,000s of SIMD threads and high throughput GPUs win Additionally: Threads managed by hardware You are not required to write code for each thread and manage them yourself Easier to increase parallelism by adding more processors So, fundamental unit of a modern GPU is a stream processor

G80 and GT200 Streaming Processor Architecture

Building a Programmable GPU The future of high throughput computing is programmable stream processing So build the architecture around the unified scalar stream processing cores GeForce 8800 GTX (G80) was the first GPU architecture built with this new paradigm

G80 Replaces The Pipeline Model Host 128 Unified Streaming Processors Input Assembler Vtx Thread Issue Geom Thread Issue Setup / Rstr / ZCull Pixel Thread Issue SP TF L1 SP SP TF L1 SP SP TF L1 SP SP TF L1 SP SP TF L1 SP SP TF L1 SP SP TF L1 SP SP TF L1 SP Thread Processor L2 L2 L2 L2 L2 L2 FB FB FB FB FB FB

GT200 Adds More Processing Power Host CPU System Memory Host Interface Input Assemble Viewport / Clip / Setup / Raster / ZCull GPU Vertex Work Distribution Geometry Work Distribution Pixel Work Distribution Compute Work Distribution Interconnection Network ROP L2 ROP L2 ROP L2 ROP L2 ROP L2 ROP L2 ROP L2 ROP L2 DRAM DRAM DRAM DRAM DRAM DRAM DRAM DRAM

8800GTX (high-end G80) 16 Stream Multiprocessors Each one contains 8 unified streaming processors 128 in total GTX280 (high-end GT200) 24 Stream Multiprocessors Each one contains 8 unified streaming processors 240 in total

Inside a Stream Multiprocessor (SM) Scalar register-based ISA Multithreaded Instruction Unit Up to 1024 concurrent threads Hardware thread scheduling In-order issue 8 SP: Thread Processors IEEE 754 32-bit floating point 32-bit and 64-bit integer 16K 32-bit registers 2 SFU: Special Function Units sin, cos, log, exp Double Precision Unit IEEE 754 64-bit floating point Fused multiply-add 16KB Shared Memory TPC I-Cache MT Issue C-Cache SP SP SP SP SFU DP SP SP SP SP SFU Shared Memory

Multiprocessor Programming Model Workloads are partitioned into blocks of threads among multiprocessors a block runs to completion a block doesn t run until resources are available Allocation of hardware resources shared memory is partitioned among blocks registers are partitioned among threads Hardware thread scheduling any thread not waiting for something can run context switching is free every cycle

Memory Hierarchy of G80 and GT200 SM can directly access device memory (video memory) Not cached Read & write GT200: 140 GB/s peak SM can access device memory via texture unit Cached Read-only, for textures and constants GT200: 48 GTexels/s peak On-chip shared memory shared among threads in an SM important for communication amongst threads provides low-latency temporary storage G80 & GT200: 16KB per SM

Performance Per Millimeter For GPU, performance == throughput Cache are limited in the memory hierarchy Strategy: hide latency with computation, not cache Heavy multithreading Switch to another group of threads when the current group is waiting for memory access Implication: need large number of threads to hide latency Occupancy: typically 128 threads/sm minimum Maximum 1024 threads/sm on GT200 (total 1024 * 24 = 24,576 threads) Strategy: Single Instruction Multiple Thread (SIMT)

SIMT Thread Execution Group 32 threads (vertices, pixels or primitives) into warps Threads in warp execute same instruction at a time Shared instruction fetch/dispatch Hardware automatically handles divergence (branches) TPC I-Cache MT Issue C-Cache SP SP SP SP Warps are the primitive unit of scheduling Pick 1 of 24 warps for each instruction slot SP SP SP SP SIMT execution is an implementation choice Shared control logic leaves more space for ALUs Largely invisible to programmer SFU DP SFU Shared Memory

Shader Branching Performance G8x/G9x/GT200 branch efficiency is 32 threads (1 warp) G80 32 pixel coherence 48 pixel coherence 16 14 PS Branching Efficiency If threads diverge, both sides of branch will execute on all 32 More efficient compared to architecture with branch efficiency of 48 threads number of coherent 4x4 tiles 12 10 8 6 4 2 0% 20% 40% 60% 80% 100% 120%

Conclusion: G80 and GT200 Streaming Processor Architecture Execute in blocks can maximally exploits data parallelism Minimize incoherent memory access Adding more ALU yields better performance Performs data processing in SIMT fashion Group 32 threads into warps Threads in warp execute same instruction at a time Thread scheduling is automatically handled by hardware Context switching is free (every cycle) Transparent scalability. Easy for programming Memory latency is covered by large number of in-flight threads Cache is mainly used for read-only memory access (texture, constants).

Stream Processing for Graphics

DX10 Graphics Pipeline (Logical View) Framebuffer Framebuffer Vertex Vertex Assembly Assembly CPU CPU GPU GPU Blending Blending Rasterizer Rasterizer Vertex Vertex Shader Shader Vertex Vertex Shader Shader Vertex Vertex Shader Shader Vertex Vertex Shader Shader Vertex Vertex Shader Shader Vertex Vertex Shader Shader Vertex Vertex Shader Shader Geometry Geometry Shader Shader Texture Texture Vertex Vertex Shader Shader Vertex Vertex Shader Shader Vertex Vertex Shader Shader Pixel Pixel Shader Shader

DX10 API Mapped To GPU A-B IA C VS FB H SO GEOM GS SHADER G TEXTURE E FB H SETUP I Unified Shading Units RASTERIZER ROP F PS D, J Fixed Function Units Vertex Shading Path Geometry Shading Path Pixel Shading Path A. Front End B. Index and Vertex Input C. Topology and Viewport Process D. Z-Cull E. Texture Fetch F. Depth Test, Alpha-Blending, AA G. Shader Cores H. Frame Buffer Controller I. Primitive Setup J. Rasterization

Texturing G80 Every 2 SMs connect to a TMU (Texture Mapping Unit) 8 TMUs in total GT200 SP SP TA TA TA TA TF TF TF TF TF TF TF TF Every 3 SMs connect to a TMU 10 TMUs in total TF Each TMU contains 4 TA (Texture Addressing) and 8 TF (Texture Filtering) L1 TA:TF = 1:2, allowing full-speed FP16 bilinear filtering and free 2xAF L2

Decoupled shader & texture pipelines G7x Architecture Math A Tex Math A Tex Math A Tex G8x and later Architecture Math A Math Tex A Math Tex A Tex Cover texture latency with thousands of threads in flight

Texture Filtering Quality 16xAF, 7900GTX vs. 8800GTX NVIDIA 7900 GTX NVIDIA 8800 GTX and after

Detail of a single ROP pixel pipeline Input Shaded Fragment Input Data Shaded Fragment Data Pixel X-Bar Interconnect MSAA CSAA INT8, FP16 & FP32 formats Support Up to 16x Multisampling AA with HDR Z Comp C Comp Color & Z Compression (2x) Multiple Render Targets (8) Z ROP C ROP Frame Buffer Partition

ROP Unit in GPU: Early Z Z-Buffer removes pixels not visible Z-Cull Early-Z Pixel Killed Z-Culling removes regions of non-visible pixels @ high rate Shader Texture Early-Z removes individual pixel quads (2x2) before they are processed Z-Check Pixel Killed Memory

3DMark05 Futuremark Next Generation Antialiasing High Quality AA with incredible performance Highest Quality AA AA with HDR Coverage Sampling AA (CSAA)

Review: Aliased Rendering Only one sample per pixel Coverage, Color, Depth and Stencil Precisely evaluated, but under-sampled

Anti-Aliasing Fundamentals Pixels in an image need to represent finite areas from the scene description, not just infinitely small sample points Ideal anti-aliasing would be a smooth convolution over the pixel area Too expensive for general scenes So we approximate by taking many samples The more samples the better

Anti-Aliasing Anti-aliasing properly accounts for the contribution of all the primitives that intersect a pixel

Supersampling Store a color and depth value for each sub-sample at every pixel. Calculate unique color and depth values at every sub-sample for every pixel of every drawn triangle Strengths Robust image quality by brute force Weaknesses Inefficient: Expensive pixel shaders and texture fetches are executed for every sub-sample; wasteful because color results within a pixel are nearly identical

Multisampling (SGI, 1993) Store a unique color and depth value for each sub-sample for each pixel, but re-use one calculated color for all color subsamples for a polygon Strengths Only one color value calculated per pixel per triangle Z and stencil evaluated precisely; interpenetrations and bulkheads handled correctly Weaknesses Memory footprint N times larger than 1x Expensive to extend to 8x quality and beyond

Motivation for CSAA Multisampling evolved from 1 2 4 samples Beyond 4 sub-samples, storage cost increases faster than the image quality improves Even more true with HDR 64b and 128b per color sub-sample! For the vast majority of edge pixels, 2 colors are enough What matters is more detailed coverage information

Coverage Sampled Antialiasing Compute and store boolean coverage at 16 sub-samples Compress the redundant color and depth/stencil information into the memory footprint and bandwidth of 4 or 8 multisamples Performance of 4xMSAA with 16x quality

CSAA Quality Levels Quality Level: 4x 8x 8xQ 16x 16xQ Texture/Shader Samples Stored Color/Z Samples 1 1 1 1 1 4 4 8 4 8 Coverage Samples 4 8 8 16 16

Half Life 2 NVIDIA 4x NVIDIA 16xQ Half Life 2 Valve Software

Shadow Edges in FEAR FEAR - Monolith

FEAR 4x FEAR - Monolith

FEAR 16x Higher Quality Object Edges FEAR - Monolith

Summary of CSAA Advantages Only need to traverse the scene once Small, fixed video memory footprint Handles inter-penetrating objects Pixel shaders are only run where color detail is needed 16x stored positional coverage 16x quality for ~4x performance

GPU beyond Graphics

A scalable parallel programming model and software environment for parallel computing Minimal extensions to familiar C/C++ environment Heterogeneous serial-parallel programming model NVIDIA Corporation 2008

The Democratization of Parallel Computing GPUs and CUDA bring parallel computing to the masses Over 100M CUDA-capable GPUs sold to date 60K CUDA developers A developer kit (i.e. GPU) costs ~$200 (for 500 GFLOPS) Data-parallel supercomputers are everywhere! CUDA makes this power accessible We re already seeing innovations in data-parallel computing Massively parallel computing has become a commodity technology! NVIDIA Corporation 2008

More about CUDA at Seminar Parallel Computing with CUDA

Questions?