Professional Documents
Culture Documents
GPUs have evolved to the point where many real world applications are easily implemented on them and run significantly faster than on multi-core systems. Future computing architectures will be hybrid systems with parallel-core GPUs working in tandem with multi-core CPUs.
Jack Dongarra
Professor, University of Tennessee; Author of Linpack
The GPU has evolved into a very flexible and powerful processor:
Its programmable using high-level languages It supports 32-bit and 64-bit floating point IEEE-754 precision It offers lots of GFLOPS:
Control
ALU ALU
CPU
GPU
The fast-growing video game industry exerts strong economic pressure that forces constant innovation
GPUs
Each NVIDIA GPU has 240 parallel cores Within each core
Floating point unit Logic unit (add, sub, mul, madd) Move, compare unit Branch unit
Graphics
GPU
(Parallel Computing)
CPU
(Sequential Computing) Data Fits in Cache Larger Data Sets
Finance
Medical
Biophysics
Numerics
Audio
Video
Imaging
CUDA Parallel Programming Architecture and Model Programming the GPU in High-Level Languages
CUDA Compiler
C Fortran
CUDA Tools
DebuggerProfiler
4 cores
SM
I-Cache MT Issue C-Cache
SP SP
Work Distribution
Tesla T10
SP SP SP SP SP SP
SFU SFU
Interconnection Network ROP L2 ROP L2 ROP L2 ROP L2 ROP L2 ROP L2 ROP L2 ROP L2
DP
DRAM DRAM DRAM
DRAM
DRAM
DRAM
DRAM
DRAM
Shared Memory
High bandwidth: Sequencing (virus scanning, genomics), sorting, database, Visual computing: Graphics, image processing, tomography, machine vision, High arithmetic intensity: Dense linear algebra, PDEs, n-body, finite difference,
Heterogeneous Programming
CUDA application = serial program executing parallel kernels, all in C
Serial C code executed by a CPU thread Parallel kernel C code executed by GPU, in threads (grouped in blocks)
Serial Code Parallel Kernel KernelA<<< nBlk, nTid >>>(args); Serial Code Parallel Kernel KernelB<<< nBlk, nTid >>>(args);
NVIDIA Corporation 2008
...
...
threadID
17
Block (2, 0)
Block (2, 1)
18
Thread Block 1
1 2 3 4 5 6 7
Thread Block N - 1
0 1 2 3 4 5 6 7
Thread Cooperation
Thread cooperation is a powerful feature of CUDA Threads can cooperate via on-chip shared memory and synchronization The on-chip shared memory within one block allows: Share memory accesses, drastic memory bandwidth reduction Share intermediate results, thus: save computation Makes algorithm porting to GPUs a lot easier (vs. GPGPU and its strict stream processor model)
Transparent Scalability
Hardware is free to schedule thread blocks on any processor
Kernels scale to any number of parallel multiprocessors
Device A Device B
Kernel grid
Block 0 Block 2 Block 4 Block 2 Block 3 Block 6 Block 7 Block 1 Block 3 Block 5 Block 4 Block 5 Block 6 Block 7
Block 0
Block 1
Block 0
Block 1
Block 2
Block 3
Block 4
Block 5
Block 6
Block 7
GPU
Interconnection Network ROP L2 ROP L2 ROP L2 ROP L2 ROP L2 ROP L2 ROP L2 ROP L2
DRAM
NVIDIA Corporation 2008
DRAM
DRAM
DRAM
DRAM
DRAM
DRAM
DRAM
Block (1, 0)
Shared Memory
Registers
Registers
Registers
Registers
Thread (0, 0)
Thread (1, 0)
Thread (0, 0)
Thread (1, 0)
Global Memory
Note: The host can read & write global memory but not shared memory
NVIDIA Corporation 2008
CUDA C Code
__global__ void saxpy_parallel(int n, float a, float *x, float *y) { int i = blockIdx.x*blockDim.x + threadIdx.x; if (i < n) y[i] = a*x[i] + y[i]; } // Invoke parallel SAXPY kernel with // 256 threads/block int nblocks = (n + 255) / 256; saxpy_parallel<<<nblocks, 256>>>(n, 2.0, x, y);
25
Compilation
Any source file containing CUDA language extensions must be compiled with nvcc NVCC is a compiler driver
Works by invoking all the necessary tools and compilers like cudacc, g++, cl, ...
That must then be compiled with the rest of the application using another tool
Or PTX object code directly
Linker
28
C for CUDA
Entry point for developers who want low-level API Shared back-end compiler and optimization technology
OpenCL
PTX
GPU
OpenCL
Hardware API - similar to OpenGL and CUDA driver API Programmer has complete access to hardware device Memory managed by programmer Generates PTX
GPU
CPU
Heterogeneous Computing
Finance
Medical
Biophysics
Numerics
Audio
Video
Imaging
Resources
NVIDIA CUDA Zone (www.nvidia.com/cuda)
SDK Manuals Papers Forum Courses
32