diff --git "a/CUDA/NVIDIA_OpenCL_ProgrammingGuide.pdf.txt" "b/CUDA/NVIDIA_OpenCL_ProgrammingGuide.pdf.txt" new file mode 100644--- /dev/null +++ "b/CUDA/NVIDIA_OpenCL_ProgrammingGuide.pdf.txt" @@ -0,0 +1,5402 @@ +OpenCL +Programming Guide +for the CUDA +Architecture + +Version 2.3 + +8/27/2009 + + + + + + + + + + + + Table of Contents + +Chapter 1. Introduction ..................................................................................... 5  + +1.1  + +1.2  + +1.3  + +1.4  + +From Graphics Processing to General-Purpose Parallel Computing ................... 5  + +CUDA™: a General-Purpose Parallel Computing Architecture ........................... 7  + +CUDA’s Scalable Programming Model ............................................................. 8  + +Document’s Structure ................................................................................... 9  + +Chapter 2. OpenCL on the CUDA Architecture ................................................. 11  + +2.1  + +CUDA Architecture ...................................................................................... 11  + +2.1.1  + +Execution Model .................................................................................. 11  + +2.1.2  Memory Model ..................................................................................... 14  + +2.2  + +Compilation ................................................................................................ 16  + +2.2.1  + +PTX..................................................................................................... 16  + +2.2.2  + +Volatile ................................................................................................ 17  + +2.3  + +Compute Capability .................................................................................... 17  + +2.4  Mode Switches ........................................................................................... 18  + +2.5  Matrix Multiplication Example ...................................................................... 18  + +Chapter 3. Performance Guidelines ................................................................. 27  + +3.1  + +Instruction Performance ............................................................................. 27  + +3.1.1  + +Instruction Throughput ........................................................................ 27  + +3.1.1.1  + +Arithmetic Instructions .................................................................. 27  + +3.1.1.2  + +Control Flow Instructions ............................................................... 29  + +3.1.1.3  + +Memory Instructions ..................................................................... 30  + +3.1.1.4  + +Synchronization Instruction ........................................................... 30  + +3.1.2  Memory Bandwidth .............................................................................. 30  + +3.1.2.1  + +Global Memory .............................................................................. 31  + +3.1.2.2  + +Local Memory ............................................................................... 38  + +3.1.2.3  + +Constant Memory .......................................................................... 38  + +3.1.2.4  + +Texture Memory ........................................................................... 38  + +3.1.2.5  + +Shared Memory ............................................................................ 39  + +ii + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + 3.1.2.6  + +Registers ...................................................................................... 46  + +3.2  + +3.3  + +NDRange ................................................................................................... 46  + +Data Transfer between Host and Device ...................................................... 47  + +3.4  Warp-Level Synchronization ........................................................................ 48  + +3.5  Overall Performance Optimization Strategies ................................................ 49  + +Appendix A. Technical Specifications .............................................................. 51  + +A.1  General Specifications ................................................................................. 51  + +A.1.1  + +Specifications for Compute Capability 1.0 .............................................. 52  + +A.1.2  + +Specifications for Compute Capability 1.1 .............................................. 53  + +A.1.3  + +Specifications for Compute Capability 1.2 .............................................. 53  + +A.1.4  + +Specifications for Compute Capability 1.3 .............................................. 53  + +A.2  + +A.3  + +Floating-Point Standard .............................................................................. 53  + +Supported OpenCL Extensions ..................................................................... 54  + +Appendix B. Mathematical Functions Accuracy ............................................... 55  + +B.1  + +Standard Functions ..................................................................................... 55  + +B.1.1  + +Single-Precision Floating-Point Functions ............................................... 55  + +B.1.2  + +Double-Precision Floating-Point Functions ............................................. 57  + +B.2  + +Native Functions ......................................................................................... 59  + +NVIDIA OpenCL Programming Guide Version 2.3 + +iii + + + + + + + + List of Figures + +Figure 1-1.  Floating-Point Operations per Second and Memory Bandwidth for the CPU + +and GPU 6  + +Figure 1-2.  The GPU Devotes More Transistors to Data Processing ............................ 7  + +Figure 1-3.  CUDA is Designed to Support Various Languages and Application + +Programming Interfaces ...................................................................................... 8  + +Figure 2-1.  Grid of Thread Blocks ........................................................................... 12  + +Figure 2-2.  Automatic Scalability ............................................................................ 13  + +Figure 2-3.  CUDA Architecture ............................................................................... 16  + +Figure 3-1.  Examples of Coalesced Global Memory Access Patterns .......................... 34  + +Figure 3-2.  Examples of Global Memory Access Patterns That Are Non-Coalesced for + +Devices of Compute Capability 1.0 or 1.1 ........................................................... 35  + +Figure 3-3.  Examples of Global Memory Access Patterns That Are Non-Coalesced for + +Devices of Compute Capability 1.0 or 1.1 ........................................................... 36  + +Figure 3-4.  Examples of Global Memory Access by Devices with Compute Capability + +1.2 and Higher .................................................................................................. 37  + +Figure 3-5.  Examples of Shared Memory Access Patterns without Bank Conflicts ..... 42  + +Figure 3-6.  Example of a Shared Memory Access Pattern without Bank Conflicts ...... 43  + +Figure 3-7.  Examples of Shared Memory Access Patterns with Bank Conflicts ........... 44  + +Figure 3-8.  Example of Shared Memory Read Access Patterns with Broadcast ........... 45  + +iv + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + + Chapter 1. +Introduction + +1.1 + +From Graphics Processing to +General-Purpose Parallel Computing + +Driven by the insatiable market demand for realtime, high-definition 3D graphics, +the programmable Graphic Processor Unit or GPU has evolved into a highly +parallel, multithreaded, manycore processor with tremendous computational +horsepower and very high memory bandwidth, as illustrated by Figure 1-1. + + + + + Chapter 1. 0BIntroduction + +GT200 + +G92 + +G80 +Ultra + +G80 + +NV35 + +NV40 + +NV30 + +G71 + +G70 + +3.0 GHz +Core2 Duo + +3.2 GHz +Harpertown + +Jan + +Jun + +Apr + +2003 + +2004 + +Jun + +2005 + +Mar + +Nov + +May + +2006 + +2007 + +Jun + +2008 + +GT200 = GeForce GTX 280 + +G71 = GeForce 7900 GTX + +NV35 = GeForce FX 5950 Ultra + +G92 = GeForce 9800 GTX + +G70 = GeForce 7800 GTX + +NV30 = GeForce FX 5800 + +G80 = GeForce 8800 GTX + +NV40 = GeForce 6800 Ultra + +G80 +Ultra + +G80 + +G71 + +NV40 + +NV30 + +Harpertown + +Woodcrest + +Prescott EE + +Northwood + +Figure 1-1. Floating-Point Operations per Second and Memory + +Bandwidth for the CPU and GPU + +The reason behind the discrepancy in floating-point capability between the CPU and +the GPU is that the GPU is specialized for compute-intensive, highly parallel +computation – exactly what graphics rendering is about – and therefore designed +such that more transistors are devoted to data processing rather than data caching +and flow control, as schematically illustrated by Figure 1-2. + +6 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + + + Chapter 1. 0BIntroduction + +Control + +Cache + +DRAM + +ALU + +ALU + +ALU + +ALU + +CPU + +DRAM + +GPU + +Figure 1-2. The GPU Devotes More Transistors to Data + +Processing + +More specifically, the GPU is especially well-suited to address problems that can be +expressed as data-parallel computations – the same program is executed on many +data elements in parallel – with high arithmetic intensity – the ratio of arithmetic +operations to memory operations. Because the same program is executed for each +data element, there is a lower requirement for sophisticated flow control; and +because it is executed on many data elements and has high arithmetic intensity, the +memory access latency can be hidden with calculations instead of big data caches. + +Data-parallel processing maps data elements to parallel processing threads. Many +applications that process large data sets can use a data-parallel programming model +to speed up the computations. In 3D rendering, large sets of pixels and vertices are +mapped to parallel threads. Similarly, image and media processing applications such +as post-processing of rendered images, video encoding and decoding, image scaling, +stereo vision, and pattern recognition can map image blocks and pixels to parallel +processing threads. In fact, many algorithms outside the field of image rendering +and processing are accelerated by data-parallel processing, from general signal +processing or physics simulation to computational finance or computational biology. + +1.2 + +CUDA™: a General-Purpose Parallel +Computing Architecture + +In November 2006, NVIDIA introduced CUDA™, a general purpose parallel +computing architecture – with a new parallel programming model and instruction +set architecture – that leverages the parallel compute engine in NVIDIA GPUs to +solve many complex computational problems in a more efficient way than on a +CPU. + +As illustrated by Figure 1-3, there are several languages and application +programming interfaces that can be used to program the CUDA architecture. + +NVIDIA OpenCL Programming Guide Version 2.3 + +7 + + + + + + + + + Chapter 1. 0BIntroduction + +Figure 1-3. CUDA is Designed to Support Various Languages + +and Application Programming Interfaces + +1.3 + +CUDA’s Scalable Programming Model + +The advent of multicore CPUs and manycore GPUs means that mainstream +processor chips are now parallel systems. Furthermore, their parallelism continues +to scale with Moore’s law. The challenge is to develop application software that +transparently scales its parallelism to leverage the increasing number of processor +cores, much as 3D graphics applications transparently scale their parallelism to +manycore GPUs with widely varying numbers of cores. + +CUDA’s parallel programming model is designed to overcome this challenge with +three key abstractions: a hierarchy of thread groups, a hierarchy of shared memories, +and barrier synchronization. + +These abstractions provide fine-grained data parallelism and thread parallelism, +nested within coarse-grained data parallelism and task parallelism. They guide the +programmer to partition the problem into coarse sub-problems that can be solved +independently in parallel, and then into finer pieces that can be solved cooperatively +in parallel. Such a decomposition preserves language expressivity by allowing +threads to cooperate when solving each sub-problem, and at the same time enables +transparent scalability since each sub-problem can be scheduled to be solved on any +of the available processor cores: A compiled program can therefore execute on any +number of processor cores, and only the runtime system needs to know the physical +processor count. + +This scalable programming model allows the CUDA architecture to span a wide +market range by simply scaling the number of processors and memory partitions: +from the high-performance enthusiast GeForce GTX 280 GPU and professional +Quadro and Tesla computing products to a variety of inexpensive, mainstream +GeForce GPUs (see Appendix A for a list of all CUDA-enabled GPUs). + +8 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + Chapter 1. 0BIntroduction + +1.4 + +Document’s Structure + +This document is organized into the following chapters: + +(cid:137) Chapter 1 is a general introduction to GPU computing and the CUDA + +architecture. + +(cid:137) Chapter 2 describes how the OpenCL architecture maps to the CUDA +architecture and the specifics of NVIDIA’s OpenCL implementation. +(cid:137) Chapter 3 gives some guidance on how to achieve maximum performance. +(cid:137) Appendix A lists the CUDA-enabled GPUs with their technical specifications. +(cid:137) Appendix B lists the accuracy of each mathematical function on the CUDA + +architecture. + +NVIDIA OpenCL Programming Guide Version 2.3 + +9 + + + + + + + + Chapter 2. +OpenCL on the CUDA Architecture + +2.1 + +CUDA Architecture + +2.1.1 + +Execution Model +The CUDA architecture is a close match to the OpenCL architecture. + +A CUDA device is built around a scalable array of multithreaded Streaming +Multiprocessors (SMs). A multiprocessor corresponds to an OpenCL compute unit. + +A multiprocessor executes a CUDA thread for each OpenCL work-item and a thread +block for each OpenCL work-group. A kernel is executed over an OpenCL +NDRange by a grid of thread blocks. As illustrated in Figure 2-1, each of the thread +blocks that execute a kernel is therefore uniquely identified by its work-group ID, +and each thread by its global ID or by a combination of its local ID and work-group +ID. + + + + + Chapter 2. 1BOpenCL on the CUDA Architecture + +Grid + +Block (0, 0) + +Block (1, 0) + +Block (2, 0) + +Block (0, 1) + +Block (1, 1) + +Block (2, 1) + +Block (1, 1) + +Thread (0, 0) Thread (1, 0) Thread (2, 0) Thread (3, 0) + +Thread (0, 1) Thread (1, 1) Thread (2, 1) Thread (3, 1) + +Thread (0, 2) Thread (1, 2) Thread (2, 2) Thread (3, 2) + +A kernel is executed over an NDRange by a grid of thread blocks. + +Figure 2-1. Grid of Thread Blocks + +A thread is also given a unique thread ID within its block. The local ID of a thread +and its thread ID relate to each other in a straightforward way: For a one- +dimensional block, they are the same; for a two-dimensional block of size (Dx, Dy), +the thread ID of a thread of index (x, y) is (x + y Dx); for a three-dimensional block +of size (Dx, Dy, Dz), the thread ID of a thread of index (x, y, z) is +(x + y Dx + z Dx Dy). + +When an OpenCL program on the host invokes a kernel, the work-groups are +enumerated and distributed as thread blocks to the multiprocessors with available +execution capacity. The threads of a thread block execute concurrently on one +multiprocessor. As thread blocks terminate, new blocks are launched on the vacated +multiprocessors. + +Thread blocks are required to execute independently: It must be possible to execute +them in any order, in parallel or in series. This independence requirement allows +thread blocks to be scheduled in any order across any number of cores, enabling + +12 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + + programmers to write code that scales with the number of cores, as illustrated in +Figure 2-2. + +Chapter 1 + +Kernel + +Block 0 + +Block 1 + +Block 2 + +Block 3 + +Block 4 + +Block 5 +Block 5 + +Block 6 +Block 6 + +Block 7 + +Device with 2 SMs + +Device with 4 SMs + +SM 0 + +SM 1 + +SM 0 + +SM 1 + +SM 2 + +SM 3 + + Block 0 + + Block 1 + + Block 0 + + Block 1 + + Block 2 + + Block 3 + + Block 4 + + Block 5 + + Block 6 + + Block 7 + + Block 2 + + Block 3 + + Block 4 + + Block 5 + + Block 6 + + Block 7 + +A device with more multiprocessors will automatically execute a kernel in less time than a device with +fewer multiprocessors. + +Figure 2-2. Automatic Scalability + +A multiprocessor consists of eight Scalar Processor (SP) cores, two special function +units for transcendentals, a multithreaded instruction unit, and on-chip shared +memory, which is used to implement OpenCL local memory. The multiprocessor +creates, manages, and executes concurrent threads in hardware with zero scheduling +overhead. It implements the work-group barrier function with a single instruction. +Fast barrier synchronization together with lightweight thread creation and zero- +overhead thread scheduling efficiently support very fine-grained parallelism, +allowing, for example, a low granularity decomposition of problems by assigning +one thread to each data element (such as a pixel in an image, a voxel in a volume, a +cell in a grid-based computation). + +To manage hundreds of threads running several different programs, the +multiprocessor employs a new architecture we call SIMT (single-instruction, +multiple-thread). The multiprocessor maps each thread to one SP, and each scalar + +NVIDIA OpenCL Programming Guide Version 2.3 + +13 + + + + + + + + + + + + + + Chapter 2. 1BOpenCL on the CUDA Architecture + +thread executes independently with its own instruction address and register state. +The multiprocessor SIMT unit creates, manages, schedules, and executes threads in +groups of 32 parallel threads called warps. (This term originates from weaving, the +first parallel thread technology. A half-warp is either the first or second half of a +warp.) Individual threads composing a SIMT warp start together at the same +program address but are otherwise free to branch and execute independently. + +When a multiprocessor is given one or more thread blocks to execute, it splits them +into warps that get scheduled by the SIMT unit. The way a block is split into warps +is always the same; each warp contains threads of consecutive, increasing thread IDs +with the first warp containing the thread of thread ID zero. + +Every instruction issue time, the SIMT unit selects a warp that is ready to execute +and issues the next instruction to the active threads of the warp. A warp executes +one common instruction at a time, so full efficiency is realized when all 32 threads +of a warp agree on their execution path. If threads of a warp diverge via a data- +dependent conditional branch, the warp serially executes each branch path taken, +disabling threads that are not on that path, and when all paths complete, the threads +converge back to the same execution path. Branch divergence occurs only within a +warp; different warps execute independently regardless of whether they are +executing common or disjointed code paths. + +SIMT architecture is akin to SIMD (Single Instruction, Multiple Data) vector +organizations in that a single instruction controls multiple processing elements. A +key difference is that SIMD vector organizations expose the SIMD width to the +software, whereas SIMT instructions specify the execution and branching behavior +of a single thread. In contrast with SIMD vector machines, SIMT enables +programmers to write thread-level parallel code for independent, scalar threads, as +well as data-parallel code for coordinated threads. For the purposes of correctness, +the programmer can essentially ignore the SIMT behavior; however, substantial +performance improvements can be realized by taking care that the code seldom +requires threads in a warp to diverge. In practice, this is analogous to the role of +cache lines in traditional code: Cache line size can be safely ignored when designing +for correctness but must be considered in the code structure when designing for +peak performance. Vector architectures, on the other hand, require the software to +coalesce loads into vectors and manage divergence manually. + +2.1.2 + +Memory Model +As illustrated by Figure 2-3, each multiprocessor has on-chip memory of the four +following types: + +(cid:137) One set of local 32-bit registers per processor, +(cid:137) A parallel data cache or shared memory that is shared by all scalar processor cores + +and is where OpenCL local memory resides, + +(cid:137) A read-only constant cache that is shared by all scalar processor cores and speeds + +up reads from OpenCL constant memory, + +(cid:137) A read-only texture cache that is shared by all scalar processor cores and speeds up +reads from OpenCL image objects; each multiprocessor accesses the texture +cache via a texture unit that implements the various addressing modes and data +filtering specified by OpenCL sampler objects; the region of device memory +addressed by image objects is referred to a texture memory. + +14 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + Chapter 1 + +There is also a global memory address space that is used for OpenCL global +memory and a local memory address space that is private to each thread (and should +not be confused with OpenCL local memory). Both memory spaces are read-write +regions of device memory and are not cached. + +A variable in OpenCL private memory generally resides in a register. However in +some cases the compiler might choose to place it in CUDA local memory, which +can have adverse performance consequences because of local memory high latency +and bandwidth (see Section 3.1.2.2). Variables that are likely to be placed in CUDA +local memory are large structures or arrays that would consume too much register +space, and arrays for which the compiler cannot determine that they are indexed +with constant quantities. + +The number of blocks a multiprocessor can process at once – referred to as the +number of active blocks per multiprocessor – depends on how many registers per +thread and how much shared memory per block are required for a given kernel since +the multiprocessor’s registers and shared memory are split among all the threads of +the active blocks. If there are not enough registers or shared memory available per +multiprocessor to process at least one block, the kernel will fail to launch. The +maximum number of active blocks per multiprocessor, as well as the maximum +number of active warps and maximum number of active threads are given in +Appendix A. + +If a non-atomic instruction executed by a warp writes to the same location in global +or shared memory for more than one of the threads of the warp, the number of +serialized writes that occur to that location and the order in which they occur is +undefined, but one of the writes is guaranteed to succeed. If an atomic instruction +executed by a warp reads, modifies, and writes to the same location in global +memory for more than one of the threads of the warp, each read, modify, write to +that location occurs and they are all serialized, but the order in which they occur is +undefined. + +NVIDIA OpenCL Programming Guide Version 2.3 + +15 + + + + + + Chapter 2. 1BOpenCL on the CUDA Architecture + +Device + +Multiprocessor N + +Multiprocessor 2 + +Multiprocessor 1 + +Shared Memory + +Registers + +Registers + +Registers + +Processor 1 + +Processor 2 + +… + +Processor M + +Instruction +Unit + +Constant +Cache + +Texture +Cache + +Device Memory + +A set of SIMT multiprocessors with on-chip shared memory. + +Figure 2-3. CUDA Architecture + +2.2 + +Compilation + +2.2.1 + +PTX +Kernels written in OpenCL C are compiled into PTX, which is CUDA’s instruction +set architecture and is described in a separate document. + +Currently, the PTX intermediate representation can be obtained by calling +clGetProgramInfo() with CL_PROGRAM_BINARIES and can be passed to + +16 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + 2.2.2 + +Chapter 1 + +clCreateProgramWithBinary() to create a program object, but this will likely +not be supported in future versions. + +Volatile +Only after the execution of barrier(), mem_fence(), read_mem_fence(), or +write_mem_fence() are prior writes to global or shared memory of a given +thread guaranteed to be visible by other threads. As long as this requirement is met, +the compiler is free to optimize reads and writes to global or shared memory. For +example, in the code sample below, the first reference to myArray[tid] compiles +into a global or shared memory read instruction, but the second reference does not +as the compiler simply reuses the result of the first read. + +// myArray is an array of non-zero integers +// located in global or shared memory +__kernel void myKernel(__global int* result) { + int tid = get_local_id(0); + int ref1 = myArray[tid] * 1; + myArray[tid + 1] = 2; + int ref2 = myArray[tid] * 1; + result[tid] = ref1 * ref2; +} +Therefore, ref2 cannot possibly be equal to 2 in thread tid as a result of thread +tid-1 overwriting myArray[tid] by 2. + +This behavior can be changed using the volatile keyword: If a variable located in +global or shared memory is declared as volatile, the compiler assumes that its value +can be changed at any time by another thread and therefore any reference to this +variable compiles to an actual memory read instruction. + +Note that even if myArray is declared as volatile in the code sample above, there is +no guarantee, in general, that ref2 will be equal to 2 in thread tid since thread +tid might read myArray[tid] into ref2 before thread tid-1 overwrites its +value by 2. Synchronization is required as mentioned in Section 3.4. + +2.3 + +Compute Capability + +The compute capability of a CUDA device is defined by a major revision number and a +minor revision number. + +Devices with the same major revision number are of the same core architecture. The +devices listed in Appendix A are all of compute capability 1.x (Their major revision +number is 1). + +The minor revision number corresponds to an incremental improvement to the core +architecture, possibly including new features. + +The technical specifications of the various compute capabilities are given in +Appendix A. + +NVIDIA OpenCL Programming Guide Version 2.3 + +17 + + + + + + Chapter 2. 1BOpenCL on the CUDA Architecture + +2.4 + +Mode Switches + +GPUs dedicate some DRAM memory to the so-called primary surface, which is used +to refresh the display device whose output is viewed by the user. When users initiate +a mode switch of the display by changing the resolution or bit depth of the display +(using NVIDIA control panel or the Display control panel on Windows), the +amount of memory needed for the primary surface changes. For example, if the user +changes the display resolution from 1280x1024x32-bit to 1600x1200x32-bit, the +system must dedicate 7.68 MB to the primary surface rather than 5.24 MB. (Full- +screen graphics applications running with anti-aliasing enabled may require much +more display memory for the primary surface.) On Windows, other events that may +initiate display mode switches include launching a full-screen DirectX application, +hitting Alt+Tab to task switch away from a full-screen DirectX application, or +hitting Ctrl+Alt+Del to lock the computer. + +If a mode switch increases the amount of memory needed for the primary surface, +the system may have to cannibalize memory allocations dedicated to OpenCL +applications. Therefore, a mode switch results in any call to the OpenCL runtime to +fail and return an invalid context error. + +2.5 + +Matrix Multiplication Example + +The following matrix multiplication example illustrates the typical data-parallel +approach used by OpenCL applications to achieve good performance on GPUs. It +also illustrates the use of OpenCL local memory that maps to shared memory on +the CUDA architecture. Shared memory is much faster than global memory as +mentioned in Section 2.1.2 and detailed in Section 3.1.2.5, so any opportunity to +replace global memory accesses by shared memory accesses should be exploited. + +The following code sample is a straightforward implementation of matrix +multiplication that does not take advantage of shared memory. Each thread reads +one row of A and one column of B and computes the corresponding element of C +as illustrated in Figure 2-4. A is therefore read B.width times from global memory +and B is read A.height times. + +// Host code + +// Matrices are stored in row-major order: +// M(row, col) = *(M.elements + row * M.width + col) +typedef struct { + int width; + int height; + cl_mem elements; +} Matrix; + +// Thread block size +#define BLOCK_SIZE 16 + +// Matrix multiplication - Host code +// Matrix dimensions are assumed to be multiples of BLOCK_SIZE +void MatMulHost(const Matrix A, const Matrix B, Matrix C, + +18 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + + + Chapter 1 + + const cl_context context, + const cl_kernel matMulKernel, + const cl_command_queue queue) +{ + // Load A and B to device memory + Matrix d_A; + d_A.width = A.width; d_A.height = A.height; + size_t size = A.width * A.height * sizeof(float); + d_A.elements = clCreateBuffer(context, + CL_MEM_READ_ONLY | CL_MEM_COPY_HOST_PTR, + size, A.elements, 0); + Matrix d_B; + d_B.width = B.width; d_B.height = B.height; + size = B.width * B.height * sizeof(float); + d_B.elements = clCreateBuffer(context, + CL_MEM_READ_ONLY | CL_MEM_COPY_HOST_PTR, + size, B.elements, 0); + + // Allocate C in device memory + Matrix d_C; + d_C.width = C.width; d_C.height = C.height; + size = C.width * C.height * sizeof(float); + d_C.elements = clCreateBuffer(context, + CL_MEM_WRITE_ONLY, size, 0, 0); + + // Invoke kernel + cl_uint i = 0; + clSetKernelArg(matMulKernel, i++, + sizeof(d_A.width), (void*)&d_A.width); + clSetKernelArg(matMulKernel, i++, + sizeof(d_A.height), (void*)&d_A.height); + clSetKernelArg(matMulKernel, i++, + sizeof(d_A.elements), (void*)&d_A.elements); + clSetKernelArg(matMulKernel, i++, + sizeof(d_B.width), (void*)&d_B.width); + clSetKernelArg(matMulKernel, i++, + sizeof(d_B.height), (void*)&d_B.height); + clSetKernelArg(matMulKernel, i++, + sizeof(d_B.elements), (void*)&d_B.elements); + clSetKernelArg(matMulKernel, i++, + sizeof(d_C.width), (void*)&d_C.width); + clSetKernelArg(matMulKernel, i++, + sizeof(d_C.height), (void*)&d_C.height); + clSetKernelArg(matMulKernel, i++, + sizeof(d_C.elements), (void*)&d_C.elements); + size_t localWorkSize[] = { BLOCK_SIZE, BLOCK_SIZE }; + size_t globalWorkSize[] = + { B.width / dimBlock.x, A.height / dimBlock.y }; + clEnqueueNDRangeKernel(queue, matMulKernel, 2, 0, + globalWorkSize, localWorkSize, + 0, 0, 0); + + // Read C from device memory + clEnqueueReadBuffer(queue, d_C.elements, CL_TRUE, 0, size, + C.elements, 0, 0, 0); + + // Free device memory + +NVIDIA OpenCL Programming Guide Version 2.3 + +19 + + + + + + + + + + Chapter 2. 1BOpenCL on the CUDA Architecture + + clReleaseMemObject(d_A.elements); + clReleaseMemObject(d_C.elements); + clReleaseMemObject(d_B.elements); +} + +// Kernel code + +// Matrices are stored in row-major order: +// M(row, col) = *(M.elements + row * M.width + col) +typedef struct { + int width; + int height; + __global float* elements; +} Matrix; + +// Thread block size +#define BLOCK_SIZE 16 + +// Matrix multiplication function called by MatMulKernel() +void MatMul(Matrix A, Matrix B, Matrix C) +{ + float Cvalue = 0; + int row = get_global_id(1); + int col = get_global_id(0); + for (int e = 0; e < A.width; ++e) + Cvalue += A.elements[row * A.width + e] + * B.elements[e * B.width + col]; + C.elements[row * C.width + col] = Cvalue; +} + +// Matrix multiplication kernel called by MatMulHost() +__kernel void MatMulKernel( + int Awidth, int Aheight, __global float* Aelements, + int Bwidth, int Bheight, __global float* Belements, + int Cwidth, int Cheight, __global float* Celements) +{ + Matrix A = { Awidth, Aheight, Aelements }; + Matrix B = { Bwidth, Bheight, Belements }; + Matrix C = { Cwidth, Cheight, Celements }; + matrixMul(A, B, C); +} + +20 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + + + + + Chapter 1 + +1 +- +h +t +d +w +B + +. + +i + +col + +0 + +B + +C + +t +h +g +i +e +h +B + +. + +t +h +g +i +e +h +A + +. + +A.width + +B.width + +0 + +A + +row + +A.height-1 + +Figure 2-4. Matrix Multipliation without Shared Memory + +The following code sample is an implementation of matrix multiplication that does +take advantage of shared memory. In this implementation, each thread block is +responsible for computing one square sub-matrix Csub of C and each thread within +the block is responsible for computing one element of Csub. As illustrated in Figure +2-5, Csub is equal to the product of two rectangular matrices: the sub-matrix of A of +dimension (A.width, block_size) that has the same line indices as Csub, and the sub- +matrix of B of dimension (block_size, A.width) that has the same column indices as +Csub. In order to fit into the device’s resources, these two rectangular matrices are +divided into as many square matrices of dimension block_size as necessary and Csub is +computed as the sum of the products of these square matrices. Each of these +products is performed by first loading the two corresponding square matrices from +global memory to shared memory with one thread loading one element of each +matrix, and then by having each thread compute one element of the product. Each +thread accumulates the result of each of these products into a register and once +done writes the result to global memory. + +NVIDIA OpenCL Programming Guide Version 2.3 + +21 + + + + + + + Chapter 2. 1BOpenCL on the CUDA Architecture + +By blocking the computation this way, we take advantage of fast shared memory +and save a lot of global memory bandwidth since A is only read (B.width / block_size) +times from global memory and B is read (A.height / block_size) times. + +The Matrix type from the previous code sample is augmented with a stride field, so +that sub-matrices can be efficiently represented with the same type. + +// Host code + +// Matrices are stored in row-major order: +// M(row, col) = *(M.elements + row * M.stride + col) +typedef struct { + int width; + int height; + int stride; + cl_mem elements; +} Matrix; + +// Thread block size +#define BLOCK_SIZE 16 + +// Matrix multiplication - Host code +// Matrix dimensions are assumed to be multiples of BLOCK_SIZE +void MatMulHost(const Matrix A, const Matrix B, Matrix C, + const cl_context context, + const cl_kernel matMulKernel, + const cl_command_queue queue) +{ + // Load A and B to device memory + Matrix d_A; + d_A.width = d_A.stride = A.width; d_A.height = A.height; + size_t size = A.width * A.height * sizeof(float); + d_A.elements = clCreateBuffer(context, + CL_MEM_READ_ONLY | CL_MEM_COPY_HOST_PTR, + size, A.elements, 0); + Matrix d_B; + d_B.width = d_B.stride = B.width; d_B.height = B.height; + size = B.width * B.height * sizeof(float); + d_B.elements = clCreateBuffer(context, + CL_MEM_READ_ONLY | CL_MEM_COPY_HOST_PTR, + size, B.elements, 0); + + // Allocate C in device memory + Matrix d_C; + d_C.width = d_C.stride = C.width; d_C.height = C.height; + size = C.width * C.height * sizeof(float); + d_C.elements = clCreateBuffer(context, + CL_MEM_WRITE_ONLY, size, 0, 0); + + // Invoke kernel + cl_uint i = 0; + clSetKernelArg(matMulKernel, i++, + sizeof(d_A.width), (void*)&d_A.width); + clSetKernelArg(matMulKernel, i++, + sizeof(d_A.height), (void*)&d_A.height); + clSetKernelArg(matMulKernel, i++, + +22 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + + + + + Chapter 1 + + sizeof(d_A.stride), (void*)&d_A.stride); + clSetKernelArg(matMulKernel, i++, + sizeof(d_A.elements), (void*)&d_A.elements); + clSetKernelArg(matMulKernel, i++, + sizeof(d_B.width), (void*)&d_B.width); + clSetKernelArg(matMulKernel, i++, + sizeof(d_B.height), (void*)&d_B.height); + clSetKernelArg(matMulKernel, i++, + sizeof(d_B. stride), (void*)&d_B.stride); + clSetKernelArg(matMulKernel, i++, + sizeof(d_B.elements), (void*)&d_B.elements); + clSetKernelArg(matMulKernel, i++, + sizeof(d_C.width), (void*)&d_C.width); + clSetKernelArg(matMulKernel, i++, + sizeof(d_C.height), (void*)&d_C.height); + clSetKernelArg(matMulKernel, i++, + sizeof(d_C.stride), (void*)&d_C.stride); + clSetKernelArg(matMulKernel, i++, + sizeof(d_C.elements), (void*)&d_C.elements); + size_t localWorkSize[] = { BLOCK_SIZE, BLOCK_SIZE }; + size_t globalWorkSize[] = + { B.width / dimBlock.x, A.height / dimBlock.y }; + clEnqueueNDRangeKernel(queue, matMulKernel, 2, 0, + globalWorkSize, localWorkSize, + 0, 0, 0); + + // Read C from device memory + clEnqueueReadBuffer(queue, d_C.elements, CL_TRUE, 0, size, + C.elements, 0, 0, 0); + + // Free device memory + clReleaseMemObject(d_A.elements); + clReleaseMemObject(d_C.elements); + clReleaseMemObject(d_B.elements); +} + +// Kernel code + +// Matrices are stored in row-major order: +// M(row, col) = *(M.elements + row * M.stride + col) +typedef struct { + int width; + int height; + int stride; + __global float* elements; +} Matrix; + +// Thread block size +#define BLOCK_SIZE 16 + +// Get a matrix element +float GetElement(const Matrix A, int row, int col) +{ + return A.elements[row * A.stride + col]; +} + +NVIDIA OpenCL Programming Guide Version 2.3 + +23 + + + + + + + + + + + + + Chapter 2. 1BOpenCL on the CUDA Architecture + +// Set a matrix element +void SetElement(Matrix A, int row, int col, float value) +{ + A.elements[row * A.stride + col] = value; +} + +// Get the BLOCK_SIZExBLOCK_SIZE sub-matrix Asub of A that is +// located col sub-matrices to the right and row sub-matrices down +// from the upper-left corner of A +Matrix GetSubMatrix(Matrix A, int row, int col) +{ + Matrix Asub; + Asub.width = BLOCK_SIZE; + Asub.height = BLOCK_SIZE; + Asub.stride = A.stride; + Asub.elements = + &A.elements[A.stride * BLOCK_SIZE * row + BLOCK_SIZE * col]; + return Asub; +} + +// Matrix multiplication function called by MatMulKernel() +void MatMul(Matrix C, Matrix A, Matrix B, + + __local float As[BLOCK_SIZE][BLOCK_SIZE], + __local float Bs[BLOCK_SIZE][BLOCK_SIZE]) +{ + // Block row and column + int blockRow = get_group_id(1); + int blockCol = get_group_id(0); + + // Each thread block computes one sub-matrix Csub of C + Matrix Csub = GetSubMatrix(C, blockRow, blockCol); + + // Each thread computes one element of Csub + // by accumulating results into Cvalue + float Cvalue = 0; + + // Thread row and column within Csub + int row = get_local_id(1); + int col = get_local_id(0); + + // Loop over all the sub-matrices of A and B that are + // required to compute Csub + // Multiply each pair of sub-matrices together + // and accumulate the results + for (int m = 0; m < (A.width / BLOCK_SIZE); ++m) { + + // Get sub-matrix Asub of A + Matrix Asub = GetSubMatrix(A, blockRow, m); + + // Get sub-matrix Bsub of B + Matrix Bsub = GetSubMatrix(B, m, blockCol); + + // Load Asub and Bsub from device memory to shared memory + // Each thread loads one element of each sub-matrix + As[row][col] = GetElement(Asub, row, col); + Bs[row][col] = GetElement(Bsub, row, col); + +24 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + + + + + + + + + + + + + + + + + + + + + Chapter 1 + + // Synchronize to make sure the sub-matrices are loaded + // before starting the computation + barrier(CLK_LOCAL_MEM_FENCE); + + // Multiply Asub and Bsub together + for (int e = 0; e < BLOCK_SIZE; ++e) + Cvalue += As[row][e] * Bs[e][col]; + + // Synchronize to make sure that the preceding + // computation is done before loading two new + // sub-matrices of A and B in the next iteration + barrier(CLK_LOCAL_MEM_FENCE); + } + + // Write Csub to device memory + // Each thread writes one element + SetElement(Csub, row, col, Cvalue); +} + +// Matrix multiplication kernel called by MatMulHost() +__kernel void matrixMulKernel( + int Cwidth, int Cheight, int Cstride, __global float* Celements, + int Awidth, int Aheight, int Astride, __global float* Aelements, + int Bwidth, int Bheight, int Bstride, __global float* Belements, + __local float As[BLOCK_SIZE][BLOCK_SIZE], + __local float Bs[BLOCK_SIZE][BLOCK_SIZE]) +{ + Matrix C = { Cwidth, Cheight, Cstride, Celements }; + Matrix A = { Awidth, Aheight, Astride, Aelements }; + Matrix B = { Bwidth, Bheight, Bstride, Belements }; + MatMul(A, B, C, As, Bs); +} + +NVIDIA OpenCL Programming Guide Version 2.3 + +25 + + + + + + + + + + + + + + + Chapter 2. 1BOpenCL on the CUDA Architecture + +B + +C + +blockCol + +1 +- +E +Z +I +S +_ +K +C +O +L +B + +0 + +col + +0 + +Csub + +row + +BLOCK_SIZE-1 + +E +Z +I +S +_ +K +C +O +L +B + +E +Z +I +S +_ +K +C +O +L +B + +E +Z +I +S +_ +K +C +O +L +B + +t +h +g +i +e +h +B + +. + +t +h +g +i +e +h +A + +. + +A + +w +o +R +k +c +o +b + +l + +BLOCK_SIZE + +BLOCK_SIZE + +BLOCK_SIZE + +A.width + +B.width + +Figure 2-5. Matrix Multipliation with Shared Memory + +26 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + + + + + + Chapter 3. +Performance Guidelines + +3.1 + +Instruction Performance + +To process an instruction for a warp of threads, a multiprocessor must: + +(cid:137) Read the instruction operands for each thread of the warp, +(cid:137) Execute the instruction, +(cid:137) Write the result for each thread of the warp. +Therefore, the effective instruction throughput depends on the nominal instruction +throughput as well as the memory latency and bandwidth. It is maximized by: + +(cid:137) Minimizing the use of instructions with low throughput (see Section 3.1.1), +(cid:137) Maximizing the use of the available memory bandwidth for each category of + +memory (see Section 3.1.2), + +(cid:137) Allowing the thread scheduler to overlap memory transactions with + +mathematical computations as much as possible, which requires that: +(cid:190) The program executed by the threads is of high arithmetic intensity, that is, + +has a high number of arithmetic operations per memory operation; + +(cid:190) There are many active threads per multiprocessor, as detailed in Section 3.2. + +3.1.1 + +3.1.1.1 + +Instruction Throughput +In this section, throughputs are given in number of operations per clock cycle per +multiprocessor. For a warp size of 32, an instruction is made of 32 operations. +Therefore, if T is the number of operations per clock cycle, the instruction +throughput is one instruction every 32/T clock cycles. + +All throughputs are for one multiprocessor. They must be multiplied by the number +of multiprocessors in the device to get throughput for the whole device. + +Arithmetic Instructions +For single-precision floating-point code, we highly recommend use of the float +type and the single-precision floating-point mathematical functions. When +compiling for devices without native double-precision floating-point support, such +as devices of compute capability 1.2 and lower, each double variable gets +converted to single-precision floating-point format (but retains its size of 64 bits) + + + + + Chapter 3. Performance Guidelines + +and double-precision floating-point arithmetic gets demoted to single-precision +floating-point arithmetic. + +We also recommend using native_* functions wherever possible and the +-cl-mad-enable build option, both of them can lead to large performance gains. + +Single-Precision Floating-Point Basic Arithmetic + +Throughput of single-precision floating-point add, multiply, and multiply-add is 8 +operations per clock cycle. + +Throughput of reciprocal is 2 operations per clock cycle. + +Throughput of single-precision floating-point division is 0.88 operations per clock +cycle, but native_divide(x, y) provides a faster version with a throughput of +1.6 operations per clock cycle. + +Single-Precision Floating-Point Square Root and Reciprocal Square +Root + +Throughput of reciprocal square root is 2 operations per clock cycle. + +Single-precision floating-point square root is implemented as a reciprocal square +root followed by a reciprocal instead of a reciprocal square root followed by a +multiplication, so that it gives correct results for 0 and infinity. Therefore, its +throughput is 1 operation per clock cycle. + +Single-Precision Floating-Point Logarithm + +Throughput of native_log(x) (see Section B.2) is 2 operations per clock cycle. + +Sine and Cosine + +Throughput of native_sin(x), native_cos(x), native_exp(x) is 1 +operation per clock cycle. + +sin(x), cos(x), tan(x), sincos(x) are much more expensive and even more +so if the absolute value of x needs to be reduced. + +More precisely, the argument reduction code comprises two code paths referred to +as the fast path and the slow path, respectively. + +The fast path is used for arguments sufficiently small in magnitude and essentially +consists of a few multiply-add operations. The slow path is used for arguments large +in magnitude, and consists of lengthy computations required to achieve correct +results over the entire argument range. + +At present, the argument reduction code for the trigonometric functions selects the +fast path for arguments whose magnitude is less than 48039.0f for the single- +precision functions, and less than 2147483648.0 for the double-precision functions. + +As the slow path requires more registers than the fast path, an attempt has been +made to reduce register pressure in the slow path by storing some intermediate +variables in CUDA local memory, which may affect performance because of local +memory high latency and bandwidth (see Section 3.1.2.2). At present, 28 bytes of +CUDA local memory are used by single-precision functions, and 44 bytes are used +by double-precision functions. However, the exact amount is subject to change. + +28 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + Chapter 3. Performance Guidelines + +Due to the lengthy computations and use of CUDA local memory in the slow path, +the trigonometric functions throughput is lower by one order of magnitude when +the slow path reduction is used as opposed to the fast path reduction. + +Integer Arithmetic + +Throughput of integer add is 8 operations per clock cycle. + +Throughput of 32-bit integer multiplication is 2 operations per clock cycle, but +mul24 provide 24-bit integer multiplication with a troughput of 8 operations per +clock cycle. On future architectures however, mul24 will be slower than 32-bit +integer multiplication, so we recommend to provide two kernels, one using mul24 +and the other using generic 32-bit integer multiplication, to be called appropriately +by the application. + +Integer division and modulo operation are particularly costly and should be avoided +if possible or replaced with bitwise operations whenever possible: If n is a power of +2, (i/n) is equivalent to (i>>log2(n)) and (i%n) is equivalent to (i&(n-1)); +the compiler will perform these conversions if n is literal. + +Comparison + +Throughput of compare, min, max is 8 operations per clock cycle. + +Bitwise Operations + +Throughput of any bitwise operation is 8 operations per clock cycle. + +Type Conversion + +Throughput of type conversion operations is 8 operations per clock cycle. + +Sometimes, the compiler must insert conversion instructions, introducing additional +execution cycles. This is the case for: + +(cid:137) Functions operating on char or short whose operands generally need to be + +converted to int, + +(cid:137) Double-precision floating-point constants (defined without any type suffix) used + +as input to single-precision floating-point computations. + +This last case can be avoided by using single-precision floating-point constants, +defined with an f suffix such as 3.141592653589793f, 1.0f, 0.5f. + +Control Flow Instructions +Any flow control instruction (if, switch, do, for, while) can significantly +impact the effective instruction throughput by causing threads of the same warp to +diverge, that is, to follow different execution paths. If this happens, the different +executions paths have to be serialized, increasing the total number of instructions +executed for this warp. When all the different execution paths have completed, the +threads converge back to the same execution path. + +To obtain best performance in cases where the control flow depends on the thread +ID, the controlling condition should be written so as to minimize the number of +divergent warps. This is possible because the distribution of the warps across the +block is deterministic as mentioned in Section 2.1.1. A trivial example is when the +controlling condition only depends on (get_local_id(0) / WSIZE) where +WSIZE is the warp size. In this case, no warp diverges since the controlling +condition is perfectly aligned with the warps. + +3.1.1.2 + +NVIDIA OpenCL Programming Guide Version 2.3 + +29 + + + + + + + + Chapter 3. Performance Guidelines + +Sometimes, the compiler may unroll loops or it may optimize out if or switch +statements by using branch predication instead, as detailed below. In these cases, no +warp can ever diverge. + +When using branch predication none of the instructions whose execution depends +on the controlling condition gets skipped. Instead, each of them is associated with a +per-thread condition code or predicate that is set to true or false based on the +controlling condition and although each of these instructions gets scheduled for +execution, only the instructions with a true predicate are actually executed. +Instructions with a false predicate do not write results, and also do not evaluate +addresses or read operands. + +The compiler replaces a branch instruction with predicated instructions only if the +number of instructions controlled by the branch condition is less or equal to a +certain threshold: If the compiler determines that the condition is likely to produce +many divergent warps, this threshold is 7, otherwise it is 4. + +3.1.1.3 + +Memory Instructions +Memory instructions include any instruction that reads from or writes to CUDA +shared, local, or global memory. + +Throughput of memory operations is 8 operations per clock cycle. When accessing +CUDA local or global memory, there are, in addition, 400 to 600 clock cycles of +memory latency. + +As an example, the throughput for the assignment operator in the following sample +code: + +__local float shared[32]; +__global float device[32]; +shared[threadIdx.x] = device[threadIdx.x]; +is 8 operations per clock cycle for the read from global memory, 8 operations per +clock cycle for the write to shared memory, but above all, there is a latency of 400 to +600 clock cycles to read data from global memory. + +Much of this global memory latency can be hidden by the thread scheduler if there +are sufficient independent arithmetic instructions that can be issued while waiting +for the global memory access to complete. + +3.1.1.4 + +Synchronization Instruction +Throughput for the barrier function is 8 operations per clock cycle in the case +where no thread has to wait for any other threads. + +3.1.2 + +Memory Bandwidth +The effective bandwidth of each memory space depends significantly on the +memory access pattern as detailed in the following sub-sections. + +Since device memory is of much higher latency and lower bandwidth than on-chip +memory, device memory accesses should be minimized. A typical programming +pattern is to stage data coming from device memory into shared memory; in other +words, to have each thread of a block: + +(cid:137) Load data from device memory to shared memory, + +30 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + Chapter 3. Performance Guidelines + +(cid:137) Synchronize with all the other threads of the block so that each thread can safely + +read shared memory locations that were written by different threads, + +(cid:137) Process the data in shared memory, +(cid:137) Synchronize again if necessary to make sure that shared memory has been + +updated with the results, + +(cid:137) Write the results back to device memory. + +3.1.2.1 + +Global Memory +Global memory is not cached, so it is all the more important to follow the right +access pattern to get maximum memory bandwidth, especially given how costly +accesses to device memory are. + +First, the device is capable of reading 4-byte, 8-byte, or 16-byte words from global +memory into registers in a single instruction. To have assignments such as: + +__global type device[32]; +type data = device[tid]; +compile to a single load instruction, type must be such that sizeof(type) is +equal to 4, 8, or 16 and variables of type type must be aligned to sizeof(type) +bytes (that is, have their address be a multiple of sizeof(type)). + +The alignment requirement is automatically fulfilled for built-in types. + +For structures, the size and alignment requirements can be enforced by the compiler +using the alignment specifiers __attribute__ ((aligned(8))) or +__attribute__ ((aligned(16))), such as + +struct { + float a; + float b; +} __attribute__ ((aligned(8))); +or + +struct { + float a; + float b; + float c; +} __attribute__ ((aligned(16))); +For structures larger than 16 bytes, the compiler generates several load instructions. +To ensure that it generates the minimum number of instructions, such structures +should be defined with __attribute__ ((aligned(16))) , such as + +struct { + float a; + float b; + float c; + float d; + float e; +} __attribute__ ((aligned(16))); +which is compiled into two 16-byte load instructions instead of five 4-byte load +instructions. + +Any address of a variable residing in global memory or returned by one of the +memory allocation routines from the driver or runtime API is always aligned to at +least 256 bytes. + +NVIDIA OpenCL Programming Guide Version 2.3 + +31 + + + + + + + + Chapter 3. Performance Guidelines + +Second, global memory bandwidth is used most efficiently when the simultaneous +memory accesses by threads in a half-warp (during the execution of a single read or +write instruction) can be coalesced into a single memory transaction of 32, 64, or 128 +bytes. + +The rest of this section describes the various requirements for memory accesses to +coalesce based on the compute capability of the device. If a half-warp fulfills these +requirements, coalescing is achieved even if the warp is divergent and some threads +of the half-warp do not actually access memory. + +For the purpose of the following discussion, global memory is considered to be +partitioned into segments of size equal to 32, 64, or 128 bytes and aligned to this +size. + +Coalescing on Devices with Compute Capability 1.0 and 1.1 + +The global memory access by all threads of a half-warp is coalesced into one or two +memory transactions if it satisfies the following three conditions: + +(cid:137) Threads must access + +(cid:137) Either 4-byte words, resulting in one 64-byte memory transaction, +(cid:137) Or 8-byte words, resulting in one 128-byte memory transaction, +(cid:137) Or 16-byte words, resulting in two 128-byte memory transactions; +(cid:137) All 16 words must lie in the same segment of size equal to the memory + +transaction size (or twice the memory transaction size when accessing 16-byte +words); + +(cid:137) Threads must access the words in sequence: The kth thread in the half-warp must + +access the kth word. + +If a half-warp does not fulfill all the requirements above, a separate memory +transaction is issued for each thread and throughput is significantly reduced. + +Figure 3-1 shows some examples of coalesced memory accesses, while Figure 3-2 +and Figure 3-3 show some examples of memory accesses that are non-coalesced for +devices of compute capability 1.0 or 1.1. + +Coalesced 8-byte accesses deliver a little lower bandwidth than coalesced 4-byte +accesses and coalesced 16-byte accesses deliver a noticeably lower bandwidth than +coalesced 4-byte accesses. But, while bandwidth for non-coalesced accesses is +around an order of magnitude lower than for coalesced accesses when these +accesses are 4-byte, it is only around four times lower when they are 8-byte and +around two times when they are 16-byte. + +Coalescing on Devices with Compute Capability 1.2 and Higher + +The global memory access by all threads of a half-warp is coalesced into a single +memory transaction as soon as the words accessed by all threads lie in the same +segment of size equal to: + +(cid:137) 32 bytes if all threads access 1-byte words, +(cid:137) 64 bytes if all threads access 2-byte words, +(cid:137) 128 bytes if all threads access 4-byte or 8-byte words. +Coalescing is achieved for any pattern of addresses requested by the half-warp, +including patterns where multiple threads access the same address. This is in + +32 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + Chapter 3. Performance Guidelines + +contrast with devices of lower compute capabilities where threads need to access +words in sequence. + +If a half-warp addresses words in n different segments, n memory transactions are +issued (one for each segment), whereas devices with lower compute capabilities +would issue 16 transactions as soon as n is greater than 1. In particular, if threads +access 16-byte words, at least two memory transactions are issued. + +Unused words in a memory transaction are still read, so they waste bandwidth. To +reduce waste, hardware will automatically issue the smallest memory transaction that +contains the requested words. For example, if all the requested words lie in one half +of a 128-byte segment, a 64-byte transaction will be issued. + +More precisely, the following protocol is used to issue a memory transaction for a +half-warp: + +(cid:137) Find the memory segment that contains the address requested by the lowest + +numbered active thread. Segment size is 32 bytes for 1-byte data, 64 bytes for +2-byte data, 128 bytes for 4-, 8- and 16-byte data. + +(cid:137) Find all other active threads whose requested address lies in the same segment. +(cid:137) Reduce the transaction size, if possible: + +(cid:137) If the transaction size is 128 bytes and only the lower or upper half is used, + +reduce the transaction size to 64 bytes; + +(cid:137) If the transaction size is 64 bytes and only the lower or upper half is used, + +reduce the transaction sizez to 32 bytes. + +(cid:137) Carry out the transaction and mark the serviced threads as inactive. +(cid:137) Repeat until all threads in the half-warp are serviced. + +Figure 3-4 shows some examples of global memory accesses for devices of compute +capability 1.2 and higher. + +NVIDIA OpenCL Programming Guide Version 2.3 + +33 + + + + + + + + + Chapter 3. Performance Guidelines + +Thread 0 + +Address 128 + +Thread 0 + +Address 128 + +Thread 1 + +Address 132 + +Thread 1 + +Address 132 + +Thread 2 + +Address 136 + +Thread 2 + +Address 136 + +Thread 3 + +Address 140 + +Thread 3 + +Address 140 + +Thread 4 + +Address 144 + +Thread 4 + +Address 144 + +Thread 5 + +Address 148 + +Thread 5 + +Address 148 + +Thread 6 + +Address 152 + +Thread 6 + +Address 152 + +Thread 7 + +Address 156 + +Thread 7 + +Address 156 + +Thread 8 + +Address 160 + +Thread 8 + +Address 160 + +Thread 9 + +Address 164 + +Thread 9 + +Address 164 + +Thread 10 + +Address 168 + +Thread 10 + +Address 168 + +Thread 11 + +Address 172 + +Thread 11 + +Address 172 + +Thread 12 + +Address 176 + +Thread 12 + +Address 176 + +Thread 13 + +Address 180 + +Thread 13 + +Address 180 + +Thread 14 + +Address 184 + +Thread 14 + +Address 184 + +Thread 15 + +Address 188 + +Thread 15 + +Address 188 + +Left: coalesced float memory access, resulting in a single memory transaction. +Right: coalesced float memory access (divergent warp), resulting in a single memory transaction. + +Figure 3-1. Examples of Coalesced Global Memory Access + +Patterns + +34 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + Chapter 3. Performance Guidelines + +Thread 0 + +Address 128 + +Thread 0 + +Address 128 + +Thread 1 + +Address 132 + +Thread 1 + +Address 132 + +Thread 2 + +Address 136 + +Thread 2 + +Address 136 + +Thread 3 + +Address 140 + +Thread 3 + +Address 140 + +Thread 4 + +Address 144 + +Thread 4 + +Address 144 + +Thread 5 + +Address 148 + +Thread 5 + +Address 148 + +Thread 6 + +Address 152 + +Thread 6 + +Address 152 + +Thread 7 + +Address 156 + +Thread 7 + +Address 156 + +Thread 8 + +Address 160 + +Thread 8 + +Address 160 + +Thread 9 + +Address 164 + +Thread 9 + +Address 164 + +Thread 10 + +Address 168 + +Thread 10 + +Address 168 + +Thread 11 + +Address 172 + +Thread 11 + +Address 172 + +Thread 12 + +Address 176 + +Thread 12 + +Address 176 + +Thread 13 + +Address 180 + +Thread 13 + +Address 180 + +Thread 14 + +Address 184 + +Thread 14 + +Address 184 + +Thread 15 + +Address 188 + +Thread 15 + +Address 188 + +Left: non-sequential float memory access, resulting in 16 memory transactions. +Right: access with a misaligned starting address, resulting in 16 memory transactions. + +Figure 3-2. Examples of Global Memory Access Patterns That + +Are Non-Coalesced for Devices of Compute +Capability 1.0 or 1.1 + +NVIDIA OpenCL Programming Guide Version 2.3 + +35 + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + Chapter 3. Performance Guidelines + +Thread 0 + +Address 128 + +Thread 0 + +Address 128 + +Thread 1 + +Address 132 + +Thread 2 + +Address 136 + +Thread 3 + +Address 140 + +Thread 1 + +Address 140 + +Thread 4 + +Address 144 + +Thread 5 + +Address 148 + +Thread 6 + +Address 152 + +Thread 2 + +Address 152 + +Thread 7 + +Address 156 + +Thread 8 + +Address 160 + +Thread 9 + +Address 164 + +Thread 3 + +Address 164 + +Thread 10 + +Address 168 + +Thread 11 + +Address 172 + +Thread 12 + +Address 176 + +Thread 4 + +Address 176 + +Thread 13 + +Address 180 + +Thread 14 + +Address 184 + +Thread 15 + +Address 188 + +Thread 5 + +Address 188 + +Left: non-contiguous float memory access, resulting in 16 memory transactions. +Right: non-coalesced float3 memory access, resulting in 16 memory transactions. + +Figure 3-3. Examples of Global Memory Access Patterns That + +Are Non-Coalesced for Devices of Compute +Capability 1.0 or 1.1 + +36 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + Chapter 3. Performance Guidelines + +6 +4 +B +s +e +g +m +e +n +t + +Thread +0 + +Thread +1 + +Thread +2 + +Thread +3 + +Thread +4 + +Thread +5 + +Thread +6 + +Thread +7 + +Thread +8 + +Thread +9 + +Thread +10 + +Thread +11 + +Thread +12 + +Thread +13 + +Thread +14 + +Thread +15 + +Thread +0 + +Thread +1 + +Thread +2 + +Thread +3 + +Thread +4 + +Thread +5 + +Thread +6 + +Thread +7 + +Thread +8 + +Thread +9 + +Thread +10 + +Thread +11 + +Thread +12 + +Thread +13 + +Thread +14 + +Thread +15 + +Address +120 + +Address +124 + +Address +128 + +Address +132 + +Address +136 + +Address +140 + +Address +144 + +Address +148 + +Address +152 + +Address +156 + +Address +160 + +Address +164 + +Address +168 + +Address +172 + +Address +176 + +Address +180 + +Address +184 + +Address +188 + +Address +192 + +Address +196 + +Address +200 + +Address +204 + +Address +208 + +Address + +212 + +Address +120 + +Address +124 + +Address +128 + +Address +132 + +Address +136 + +Address +140 + +Address +144 + +Address +148 + +Address +152 + +Address +156 + +Address +160 + +Address +164 + +Address +168 + +Address +172 + +Address +176 + +Address +180 + +Address +184 + +Address +188 + +Address +192 + +Address +19 6 + +… + +Address +204 + +Address +252 + +Address + +256 + +1 +2 +8 +B +s +e +g +m +e +n +t + +Thread +0 + +Thread +1 + +Thread +2 + +Thread +3 + +Thread +4 + +Thread +5 + +Thread +6 + +Thread +7 + +Thread +8 + +Thread +9 + +Thread +10 + +Thread +11 + +Thread +12 + +Thread +13 + +Thread +14 + +Thread +15 + +3 +2 +B +s +e +g +m +e +n +t + +6 +4 +B +s +e +g +m +e +n +t + +Address +96 + +Address +100 + +Address +104 + +Address +108 + +Address +112 + +Address +116 + +Address +120 + +Address +124 + +Address +128 + +Address +132 + +Address +136 + +Address +140 + +Address +144 + +Address +148 + +Address +152 + +Address +156 + +Address +160 + +Address +164 + +Address +168 + +Address +172 + +Address +176 + +Address +180 + +Address +184 + +Address + +188 + +Left: random float memory access within a 64B segment, resulting in one memory transaction. +Center: misaligned float memory access, resulting in one transaction. +Right: misaligned float memory access, resulting in two transactions. + +Figure 3-4. Examples of Global Memory Access by Devices + +with Compute Capability 1.2 and Higher + +NVIDIA OpenCL Programming Guide Version 2.3 + +37 + + + + + + + + + + + + + + + + Chapter 3. Performance Guidelines + +Common Access Patterns + +Array of Structures + +A common global memory access pattern is when each thread of thread ID tid +accesses one element of an array located at address BaseAddress of type type* +using the following address: + + BaseAddress + tid +To get memory coalescing, type must meet the size and alignment requirements +discussed above. In particular, this means that if type is a structure larger than 16 +bytes, it should be split into several structures that meet these requirements and the +data should be laid out in memory as a list of several arrays of these structures +instead of a single array of type type*. + +Two-Dimensional Array + +Another common global memory access pattern is when each thread of index +(tx,ty) accesses one element of a 2D array located at address BaseAddress of +type type* and of width width using the following address: + + BaseAddress + width * ty + tx +In such a case, one gets memory coalescing for all half-warps of the thread block +only if: + +(cid:137) The width of the thread block is a multiple of half the warp size; +(cid:137) width is a multiple of 16. +In particular, this means that an array whose width is not a multiple of 16 will be +accessed much more efficiently if it is actually allocated with a width rounded up to +the closest multiple of 16 and its rows padded accordingly. + +Local Memory +Like global memory, CUDA local memory is not cached, so accesses to local +memory are as expensive as accesses to global memory. Local memory accesses are +always coalesced though since they are per-thread by definition. + +Constant Memory +Constant memory is cached so a read from constant memory costs one memory +read from device memory only on a cache miss, otherwise it just costs one read +from the constant cache. + +For all threads of a half-warp, reading from the constant cache is as fast as reading +from a register as long as all threads read the same address. The cost scales linearly +with the number of different addresses read by all threads. We recommend having +all threads of the entire warp read the same address as opposed to all threads within +each of its halves only, as future devices will require it for full speed read. + +Texture Memory +Texture memory is cached so an image read costs one memory read from device +memory only on a cache miss, otherwise it just costs one read from the texture +cache. The texture cache is optimized for 2D spatial locality, so threads of the same +warp that read image addresses that are close together will achieve best +performance. Also, it is designed for streaming reads with a constant latency, i.e. a +cache hit reduces DRAM bandwidth demand, but not read latency. + +3.1.2.2 + +3.1.2.3 + +3.1.2.4 + +38 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + Chapter 3. Performance Guidelines + +Reading device memory through image objects present some benefits that can make +it an advantageous alternative to reading device memory from global or constant +memory: + +(cid:137) If the memory reads do not follow the access patterns that global or constant +memory reads must respect to get good performance (see Sections 3.1.2.1 and +3.1.2.3), higher bandwidth can be achieved providing that there is locality in the +image reads; + +(cid:137) The latency of addressing calculations is hidden better, possibly improving +performance for applications that perform random accesses to the data; +(cid:137) Packed data may be broadcast to separate variables in a single operation; +(cid:137) 8-bit and 16-bit integer input data may be optionally converted to 32-bit floating- + +point values in the range [0.0, 1.0] or [-1.0, 1.0]. + +However, within the same kernel call, the texture cache is not kept coherent with +respect to image writes, so that any image read to an address that has been written +to via an image write in the same kernel call returns undefined data. In other words, +a thread can safely read via an image object some memory location only if this +memory location has been updated by a previous kernel call or memory copy, but +not if it has been previously updated by the same thread or another thread from the +same kernel call. + +3.1.2.5 + +Shared Memory +Shared memory is where OpenCL local memory resides. + +Because it is on-chip, shared memory is much faster than local and global memory. +In fact, for all threads of a warp, accessing shared memory is as fast as accessing a +register as long as there are no bank conflicts between the threads, as detailed below. + +To achieve high memory bandwidth, shared memory is divided into equally-sized +memory modules, called banks, which can be accessed simultaneously. So, any +memory read or write request made of n addresses that fall in n distinct memory +banks can be serviced simultaneously, yielding an effective bandwidth that is n times +as high as the bandwidth of a single module. + +However, if two addresses of a memory request fall in the same memory bank, there +is a bank conflict and the access has to be serialized. The hardware splits a memory +request with bank conflicts into as many separate conflict-free requests as necessary, +decreasing the effective bandwidth by a factor equal to the number of separate +memory requests. If the number of separate memory requests is n, the initial +memory request is said to cause n-way bank conflicts. + +To get maximum performance, it is therefore important to understand how memory +addresses map to memory banks in order to schedule the memory requests so as to +minimize bank conflicts. + +In the case of shared memory, the banks are organized such that successive 32-bit +words are assigned to successive banks and each bank has a bandwidth of 32 bits +per two clock cycles. + +For devices of compute capability 1.x, the warp size is 32 and the number of banks +is 16 (see Section 5.1); a shared memory request for a warp is split into one request +for the first half of the warp and one request for the second half of the warp. As a +consequence, there can be no bank conflict between a thread belonging to the first +half of a warp and a thread belonging to the second half of the same warp. + +NVIDIA OpenCL Programming Guide Version 2.3 + +39 + + + + + + + + Chapter 3. Performance Guidelines + +A common case is for each thread to access a 32-bit word from an array indexed by +the thread ID tid and with some stride s: + +__local float shared[32]; +float data = shared[BaseIndex + s * tid]; +In this case, the threads tid and tid+n access the same bank whenever s*n is a +multiple of the number of banks m or equivalently, whenever n is a multiple of m/d +where d is the greatest common divisor of m and s. As a consequence, there will be +no bank conflict only if half the warp size is less than or equal to m/d. For devices +of compute capability 1.x, this translates to no bank conflict only if d is equal to 1, +or in other words, only if s is odd since m is a power of two. + +Figure 3-5 and Figure 3-6 show some examples of conflict-free memory accesses +while Figure 3-7 shows some examples of memory accesses that cause bank +conflicts. + +Other cases worth mentioning are when each thread accesses an element that is +smaller or larger than 32 bits in size. For example, there are bank conflicts if an array +of char is accessed the following way: + +__local char shared[32]; +char data = shared[BaseIndex + tid]; +because shared[0], shared[1], shared[2], and shared[3], for example, +belong to the same bank. There are no bank conflicts however, if the same array is +accessed the following way: + +char data = shared[BaseIndex + 4 * tid]; +There are also 2-way bank conflicts for arrays of double: + +__local double shared[32]; +double data = shared[BaseIndex + tid]; +since the memory request is compiled into two separate 32-bit requests. One way to +avoid bank conflicts in this case is two split the double operands like in the +following sample code: + +__local int shared_lo[32]; +__local int shared_hi[32]; + +double dataIn; +shared_lo[BaseIndex + tid] = __double2loint(dataIn); +shared_hi[BaseIndex + tid] = __double2hiint(dataIn); + +double dataOut = + __hiloint2double(shared_hi[BaseIndex + tid], + shared_lo[BaseIndex + tid]); +It might not always improve performance though and will perform worse on future +architectures. + +A structure assignment is compiled into as many memory requests as necessary for +each member in the structure, so the following code, for example: + +__local struct type shared[32]; +struct type data = shared[BaseIndex + tid]; +results in: + +(cid:137) Three separate memory reads without bank conflicts if type is defined as +struct type { + +40 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + + Chapter 3. Performance Guidelines + +float x, y, z; + +}; + +since each member is accessed with a stride of three 32-bit words; + +(cid:137) Two separate memory reads with bank conflicts if type is defined as +struct type { + +float x, y; + +}; + +since each member is accessed with a stride of two 32-bit words; + +(cid:137) Two separate memory reads with bank conflicts if type is defined as +struct type { + +float f; +char c; + +}; + +since each member is accessed with a stride of five bytes. + +Finally, shared memory also features a broadcast mechanism whereby a 32-bit word +can be read and broadcast to several threads simultaneously when servicing one +memory read request. This reduces the number of bank conflicts when several +threads of a half-warp read from an address within the same 32-bit word. More +precisely, a memory read request made of several addresses is serviced in several +steps over time – one step every two clock cycles – by servicing one conflict-free +subset of these addresses per step until all addresses have been serviced; at each +step, the subset is built from the remaining addresses that have yet to be serviced +using the following procedure: + +(cid:137) Select one of the words pointed to by the remaining addresses as the broadcast + +word, + +(cid:137) Include in the subset: + +(cid:137) All addresses that are within the broadcast word, +(cid:137) One address for each bank pointed to by the remaining addresses. + +Which word is selected as the broadcast word and which address is picked up for +each bank at each cycle are unspecified. + +A common conflict-free case is when all threads of a half-warp read from an address +within the same 32-bit word. + +Figure 3-8 shows some examples of memory read accesses that involve the +broadcast mechanism. + +NVIDIA OpenCL Programming Guide Version 2.3 + +41 + + + + + + + + + + + + Chapter 3. Performance Guidelines + +Thread 0 + +Bank 0 + +Thread 0 + +Bank 0 + +Thread 1 + +Bank 1 + +Thread 1 + +Bank 1 + +Thread 2 + +Bank 2 + +Thread 2 + +Bank 2 + +Thread 3 + +Bank 3 + +Thread 3 + +Bank 3 + +Thread 4 + +Bank 4 + +Thread 4 + +Bank 4 + +Thread 5 + +Bank 5 + +Thread 5 + +Bank 5 + +Thread 6 + +Bank 6 + +Thread 6 + +Bank 6 + +Thread 7 + +Bank 7 + +Thread 7 + +Bank 7 + +Thread 8 + +Bank 8 + +Thread 8 + +Bank 8 + +Thread 9 + +Bank 9 + +Thread 9 + +Bank 9 + +Thread 10 + +Bank 10 + +Thread 10 + +Bank 10 + +Thread 11 + +Bank 11 + +Thread 11 + +Bank 11 + +Thread 12 + +Bank 12 + +Thread 12 + +Bank 12 + +Thread 13 + +Bank 13 + +Thread 13 + +Bank 13 + +Thread 14 + +Bank 14 + +Thread 14 + +Bank 14 + +Thread 15 + +Bank 15 + +Thread 15 + +Bank 15 + +Left: linear addressing with a stride of one 32-bit word. +Right: random permutation. + +Figure 3-5. Examples of Shared Memory Access Patterns + +without Bank Conflicts + +42 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + + Chapter 3. Performance Guidelines + +Thread 0 + +Bank 0 + +Thread 1 + +Bank 1 + +Thread 2 + +Bank 2 + +Thread 3 + +Bank 3 + +Thread 4 + +Bank 4 + +Thread 5 + +Bank 5 + +Thread 6 + +Bank 6 + +Thread 7 + +Bank 7 + +Thread 8 + +Bank 8 + +Thread 9 + +Bank 9 + +Thread 10 + +Bank 10 + +Thread 11 + +Bank 11 + +Thread 12 + +Bank 12 + +Thread 13 + +Bank 13 + +Thread 14 + +Bank 14 + +Thread 15 + +Bank 15 + +Linear addressing with a stride of three 32-bit words. + +Figure 3-6. Example of a Shared Memory Access Pattern + +without Bank Conflicts + +NVIDIA OpenCL Programming Guide Version 2.3 + +43 + + + + + + + + + + + Chapter 3. Performance Guidelines + +Thread 0 + +Bank 0 + +Thread 0 + +Bank 0 + +Thread 1 + +Bank 1 + +Thread 1 + +Bank 1 + +Thread 2 + +Bank 2 + +Thread 2 + +Bank 2 + +Thread 3 + +Bank 3 + +Thread 3 + +Bank 3 + +Thread 4 + +Bank 4 + +Thread 4 + +Bank 4 + +Thread 5 + +Bank 5 + +Thread 5 + +Bank 5 + +Thread 6 + +Bank 6 + +Thread 6 + +Bank 6 + +Thread 7 + +Bank 7 + +Thread 7 + +Bank 7 + +Thread 8 + +Bank 8 + +Thread 8 + +Bank 8 + +Thread 9 + +Bank 9 + +Thread 9 + +Bank 9 + +Thread 10 + +Bank 10 + +Thread 10 + +Bank 10 + +Thread 11 + +Bank 11 + +Thread 11 + +Bank 11 + +Thread 12 + +Bank 12 + +Thread 12 + +Bank 12 + +Thread 13 + +Bank 13 + +Thread 13 + +Bank 13 + +Thread 14 + +Bank 14 + +Thread 14 + +Bank 14 + +Thread 15 + +Bank 15 + +Thread 15 + +Bank 15 + +Left: Linear addressing with a stride of two 32-bit words causes 2-way bank conflicts. +Right: Linear addressing with a stride of eight 32-bit words causes 8-way bank conflicts. + +Figure 3-7. Examples of Shared Memory Access Patterns with + +Bank Conflicts + +44 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + + Chapter 3. Performance Guidelines + +Thread 0 + +Bank 0 + +Thread 0 + +Bank 0 + +Thread 1 + +Bank 1 + +Thread 1 + +Bank 1 + +Thread 2 + +Bank 2 + +Thread 2 + +Bank 2 + +Thread 3 + +Bank 3 + +Thread 3 + +Bank 3 + +Thread 4 + +Bank 4 + +Thread 4 + +Bank 4 + +Thread 5 + +Bank 5 + +Thread 5 + +Bank 5 + +Thread 6 + +Bank 6 + +Thread 6 + +Bank 6 + +Thread 7 + +Bank 7 + +Thread 7 + +Bank 7 + +Thread 8 + +Bank 8 + +Thread 8 + +Bank 8 + +Thread 9 + +Bank 9 + +Thread 9 + +Bank 9 + +Thread 10 + +Bank 10 + +Thread 10 + +Bank 10 + +Thread 11 + +Bank 11 + +Thread 11 + +Bank 11 + +Thread 12 + +Bank 12 + +Thread 12 + +Bank 12 + +Thread 13 + +Bank 13 + +Thread 13 + +Bank 13 + +Thread 14 + +Bank 14 + +Thread 14 + +Bank 14 + +Thread 15 + +Bank 15 + +Thread 15 + +Bank 15 + +Left: This access pattern is conflict-free since all threads read from an address within the same 32-bit +word. +Right: This access pattern causes either no bank conflicts if the word from bank 5 is chosen as the +broadcast word during the first step or 2-way bank conflicts, otherwise. + +Figure 3-8. Example of Shared Memory Read Access Patterns + +with Broadcast + +NVIDIA OpenCL Programming Guide Version 2.3 + +45 + + + + + + + + + Chapter 3. Performance Guidelines + +3.1.2.6 + +Registers +Generally, accessing a register is zero extra clock cycles per instruction, but delays +may occur due to register read-after-write dependencies and register memory bank +conflicts. + +The delays introduced by read-after-write dependencies can be ignored as soon as +there are at least 192 active threads per multiprocessor to hide them. + +The compiler and thread scheduler schedule the instructions as optimally as possible +to avoid register memory bank conflicts. They achieve best results when the number +of threads per block is a multiple of 64. Other than following this rule, an +application has no direct control over these bank conflicts. In particular, there is no +need to pack data into float4 or int4 types. + +3.2 + +NDRange + +How the NDRange affects the execution time of a kernel launch generally depends +on the kernel code. Experimentation is therefore recommended and applications +should set the work-group size explicitly as opposed to rely on the OpenCL +implementation to determine the right size (by setting local_work_size to NULL in +clEnqueueNDRangeKernel()). There are however general guidelines, described +in this section. + +For a start, the kernel will simply fail to launch if the number of threads per block +either is above the maximum number of threads per block as specified in +Appendix A, or requires too many registers or shared memory than available per +multiprocessor as mentioned in Section 2.1.2. The total number of registers required +for a block is equal to + +ceil + +( + +R + +× + +ceil + +),32,( +T + +maxR +32 + +) + +) + +R + +ceil + +yx +,( + +maxR + + is the number of registers required for the kernel, + +is equal to x rounded up to the nearest multiple of + +where +is the number of +registers per multiprocessor given in Appendix A, T is the number of threads per +. The total +block, and +amount of shared memory required for a block is equal to the sum of the amount of +statically allocated shared memory, the amount of dynamically allocated shared +memory, and the amount of shared memory used to pass the kernel’s arguments. +Note that each double or long long variable uses two registers. However, +devices of compute capability 1.2 and higher have twice as many registers per +multiprocessor as devices with lower compute capability. + +y + +Then, given a total number of threads per grid, the number of threads per block +might be dictated by the need to have enough blocks in the grid to maximize the +utilization of the available computing resources. First, there should be at least as +many blocks as there are multiprocessors in the device. Then, running only one +block per multiprocessor will force the multiprocessor to idle during thread +synchronization and also during device memory reads if there are not enough +threads per block to cover the load latency. It is therefore usually better to allow for +two or more blocks to be active on each multiprocessor to allow overlap between +blocks that wait and blocks that can run. For this to happen, not only should there +be at least twice as many blocks as there are multiprocessors in the device, but also + +46 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + Chapter 3. Performance Guidelines + +the amount of registers and shared memory required per block must be low enough +to allow for more than one active block (see Section 2.1.2). More thread blocks +stream in pipeline fashion through the device and amortize overhead even more. +The number of blocks per grid should be at least 100 if one wants it to scale to +future devices; 1000 blocks will scale across several generations. + +With a high enough number of blocks, the number of threads per block should be +chosen as a multiple of the warp size to avoid swasting computing resources with +under-populated warps, or better, a multiple of 64 for the reason invoked in +Section 3.1.2.6. Allocating more threads per block is better for efficient time slicing, +but the more threads per block, the fewer registers are available per thread, which +might prevent the kernel invocation from succeeding. + +Usually, 64 threads per block is minimal and makes sense only if there are multiple +active blocks per multiprocessor; 192 or 256 threads per block is better and usually +allows for enough registers to compile. + +The ratio of the number of active warps per multiprocessor to the maximum +number of active warps (given in Appendix A) is called the multiprocessor occupancy. +In order to maximize occupancy, the compiler attempts to minimize register usage +while keeping the number of instructions and CUDA local memory usage to a +minimum. The CUDA Software Development Kit provides a spreadsheet to assist +programmers in choosing thread block size based on shared memory and register +requirements. + +3.3 + +Data Transfer between Host and Device + +The bandwidth between device memory and the device is much higher than the +bandwidth between device memory and host memory. Therefore, one should strive +to minimize data transfer between the host and the device, for example, by moving +more code from the host to the device, even if that means running kernels with low +parallelism computations. Intermediate data structures may be created in device +memory, operated on by the device, and destroyed without ever being mapped by +the host or copied to host memory. + +Also, because of the overhead associated with each transfer, batching many small +transfers into a big one always performs much better than making each transfer +separately. + +Finally, higher performance for data transfers between host and device is achieved +for memory objects allocated in page-locked (also known as pinned) host memory (as +opposed to regular pageable host memory allocated by malloc()), which has +several benefits: + +(cid:137) Bandwidth between host memory and device memory is higher if host memory + +is allocated as page-locked. + +(cid:137) For some devices, copies between page-locked host memory and device memory + +can be performed concurrently with kernel execution. + +(cid:137) For some devices, page-locked host memory can be mapped into the device’s + +address space. In this case, there is no need to allocate any device memory and to +explicitly copy data between device and host memory. Data transfers are +implicitly performed each time the kernel accesses the mapped memory. For +maximum performance, these memory accesses must be coalesced like if they + +NVIDIA OpenCL Programming Guide Version 2.3 + +47 + + + + + + + + Chapter 3. Performance Guidelines + +were accesses to global memory (see Section 3.1.2.1). Assuming that they are an +that the mapped memory is read or written only once, avoiding explicit copies +between device and host memory can be a win performance-wise. It is always a +win on integrated systems where device memory and host memory are physically +the same and therefore any copy between host and device memory is +superfluous. +enCL applicat + +ions do not have direct control over whether memory objects are + +Op +allocated in page-locked memory or not, but they can create objects using the +CL_MEM_ALLOC_HOST_PTR flag and such objects are likely to be allocated in +locked memory by the driver for best performance. + + page- + +d + +3.4 + +Warp-Level Synchronization + +Because a warp executes one common instruction at a time, threads within a warp +are implicitly synchronized and this can be used to omit calls to the barrier() +function for better performance. + +example, both calls to barrier() are required to + +In the following code sample, for +get the expected result (i.e. result[i] = 2 * myArray[i] for i > 0). +Without synchronization, any of the two references to myArray[tid] could +return either 2 or the value initially stored in myArray, depending on whether t +memory read occurs before or after the memory write from +myArray[tid + 1] = 2. + +he + +// myArray is an array of integers located in global or shared +// memory +__kernel void myKernel(__global int* result) { + int tid = get_local_id(0); + ... + int ref1 = myArray[tid] * 1; + barrier(CLK_LOCAL_MEM_FENCE|CLK_GLOBAL_MEM_FENCE); + myArray[tid + 1] = 2; + barrier(CLK_LOCAL_MEM_FENCE|CLK_GLOBAL_MEM_FENCE); + int ref2 = myArray[tid] * 1; + result[tid] = ref1 * ref2; + ... +} +However, in the following slightly modified code sample, threads are guaranteed to +belong to the same warp, so that there is no need for any barrier() call. + +// myArray is an array of integers located in global or shared +// memory +__kernel void myKernel(__global int* result) { + int tid = get_local_id(0); + ... + if (tid < warpSize) { + int ref1 = myArray[tid] * 1; + myArray[tid + 1] = 2; + int ref2 = myArray[tid] * 1; + result[tid] = ref1 * ref2; + } + ... +} + +48 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + + + Chapter 3. Performance Guidelines + +Simply removing the call to barrier() is not enough however; myArray also +needs to be declared as volatile as described in Section 2.2.2. + +3.5 + +Overall Performance Optimization Strategies + +Performance optimization revolves around three basic strategies: + +(cid:137) Maximizing parallel execution; +(cid:137) Optimizing memory usage to achieve maximum memory bandwidth; +(cid:137) Optimizing instruction usage to achieve maximum instruction throughput. +Maximizing parallel execution starts with structuring the algorithm in a way that +exposes as much data parallelism as possible. At points in the algorithm where +parallelism is broken because some threads need to synchronize in order to share +data between each other, there are two cases: Either these threads belong to the +same block, in which case they should use the barrier() function and share data +through shared memory within the same kernel call, or they belong to different +blocks, in which case they must share data through global memory using two +separate kernel invocations, one for writing to and one for reading from global +memory. + +Once the parallelism of the algorithm has been exposed it needs to be mapped to +the hardware as efficiently as possible. This is done by carefully choosing the +NDRange of each kernel invocation as detailed in Section 3.2. + +The application should also maximize parallel execution at a higher level by +explicitly exposing concurrent execution on the device through queues, as well as +maximizing concurrent execution between host and device. + +Optimizing memory usage starts with minimizing data transfers with low- +bandwidth. That means minimizing data transfers between the host and the device, +as detailed in Section 3.3, since these have much lower bandwidth than data +transfers between device and global memory. That also means minimizing data +transfers between device and global memory by maximizing use of shared memory +on the device, as mentioned in Section 3.1.2. Sometimes, the best optimization +might even be to avoid any data transfer in the first place by simply recomputing the +data instead whenever it is needed. + +As detailed in Sections 3.1.2.1, 3.1.2.3, 3.1.2.4, and 3.1.2.5, the effective bandwidth +can vary by an order of magnitude depending on access pattern for each type of +memory. The next step in optimizing memory usage is therefore to organize +memory accesses as optimally as possible based on the optimal memory access +patterns. This optimization is especially important for global memory accesses as +global memory bandwidth is low and its latency is hundreds of clock cycles (see +Section 3.1.1.3). Shared memory accesses, on the other hand, are usually worth +optimizing only in case they have a high degree of bank conflicts. + +As for optimizing instruction usage, the use of arithmetic instructions with low +throughput (see Section 3.1.1.1) should be minimized. This includes trading +precision for speed when it does not affect the end result, such as using intrinsic +instead of regular functions (intrinsic functions are listed in Section B.2) or single- +precision instead of double-precision. Particular attention must be paid to control +flow instructions due to the SIMT nature of the device as detailed in Section 3.1.1.2. + +NVIDIA OpenCL Programming Guide Version 2.3 + +49 + + + + + + + + + Appendix A. +Technical Specifications + +A.1 + +General Specifications + +The general specifications and features of a compute device depend on its compute +capability (see Section 2.3). + +The following sections describe the technical specifications and features associated +to each compute capability. The specifications for a given compute capability are the +same as for the compute capability just below unless otherwise mentioned. Similarly, +any feature supported for a given compute capability is supported for any higher +compute capability. + +The compute capability and number of multiprocessors of all CUDA-enabled +devices are given in the following table: + +Number of +Multiprocessors + +Compute +Capability + +GeForce GTX 295 + +GeForce GTX 285, GTX 280 + +GeForce GTX 260 + +GeForce 9800 GX2 + +GeForce GTS 250, GTS 150, 9800 GTX, +9800 GTX+, 8800 GTS 512 + +GeForce 8800 Ultra, 8800 GTX + +GeForce 9800 GT, 8800 GT, 9800M GTX + +GeForce GT 130, 9600 GSO, 8800 GS, +8800M GTX, 9800M GT + +GeForce 8800 GTS + +GeForce 9600 GT, 8800M GTS, 9800M GTS + +GeForce 9700M GT + +GeForce GT 120, 9500 GT, 8600 GTS, 8600 GT, +9700M GT, 9650M GS, 9600M GT, 9600M GS, +9500M GS, 8700M GT, 8600M GT, 8600M GS + +GeForce G100, 8500 GT, 8400 GS, 8400M GT, +9500M G, 9300M G, 8400M GS, 9400 mGPU, +9300 mGPU, 8300 mGPU, 8200 mGPU, + + (1 Multiprocessor += 8 Processors) +2x30 + +30 + +24 + +2x16 + +16 + +16 + +14 + +12 + +12 + +8 + +6 + +4 + +2 + +1.3 + +1.3 + +1.3 + +1.1 + +1.1 + +1.0 + +1.1 + +1.1 + +1.0 + +1.1 + +1.1 + +1.1 + +1.1 + + + + + + + + + + Appendix A. Technical Specifications + +8100 mGPU + +GeForce 9300M GS, 9200M GS, 9100M G, +8400M G + +Tesla S1070 + +Tesla C1060 + +Tesla S870 + +Tesla D870 + +Tesla C870 + +Quadro Plex 2200 D2 + +Quadro Plex 2100 D4 + +Quadro Plex 2100 Model S4 + +Quadro Plex 1000 Model IV + +Quadro FX 5800 + +Quadro FX 4800 + +Quadro FX 4700 X2 + +Quadro FX 3700M + +Quadro FX 5600 + +Quadro FX 3700 + +Quadro FX 3600M + +Quadro FX 4600 + +Quadro FX 2700M + +Quadro FX 1700, FX 570, NVS 320M, FX 1700M, +FX 1600M, FX 770M, FX 570M + +Quadro FX 370, NVS 290, NVS 140M, NVS 135M, +FX 360M + +Quadro FX 370M, NVS 130M + +1 + +4x30 + +30 + +4x16 + +2x16 + +16 + +2x30 + +4x14 + +4x16 + +2x16 + +30 + +24 + +2x14 + +16 + +16 + +14 + +12 + +12 + +6 + +4 + +2 + +1 + +1.1 + +1.3 + +1.3 + +1.0 + +1.0 + +1.0 + +1.3 + +1.1 + +1.0 + +1.0 + +1.3 + +1.3 + +1.1 + +1.1 + +1.0 + +1.1 + +1.1 + +1.0 + +1.1 + +1.1 + +1.1 + +1.1 + +The number of multiprocessors, the clock frequency and the total amount of device +memory can be queried using the runtime. + +A.1.1 + +Specifications for Compute Capability 1.0 +(cid:137) The maximum number of threads per block is 512; +(cid:137) The maximum sizes of the x-, y-, and z-dimension of a thread block are 512, 512, + +and 64, respectively; + +(cid:137) The maximum size of each dimension of a grid of thread blocks is 65535; +(cid:137) The warp size is 32 threads; +(cid:137) The number of 32-bit registers per multiprocessor is 8192; +(cid:137) The amount of shared memory available per multiprocessor is 16 KB organized + +into 16 banks (see Section 3.1.2.5); + +(cid:137) The total amount of constant memory is 64 KB; +(cid:137) The total amount of local memory per thread is 16 KB; +(cid:137) The cache working set for constant memory is 8 KB per multiprocessor; +(cid:137) The cache working set for texture memory varies between 6 and 8 KB per + +multiprocessor; + +52 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + Appendix A. Technical Specifications + +(cid:137) The maximum number of active blocks per multiprocessor is 8; +(cid:137) The maximum number of active warps per multiprocessor is 24; +(cid:137) The maximum number of active threads per multiprocessor is 768; +(cid:137) The limit on kernel size is 2 million PTX instructions; + +A.1.2 + +A.1.3 + +Specifications for Compute Capability 1.1 +(cid:137) Support for atomic functions operating on 32-bit words in global memory. + +Specifications for Compute Capability 1.2 +(cid:137) Support for atomic functions operating in shared memory and atomic functions + +operating on 64-bit words in global memory; + +(cid:137) Support for warp vote functions; +(cid:137) The number of registers per multiprocessor is 16384; +(cid:137) The maximum number of active warps per multiprocessor is 32; +(cid:137) The maximum number of active threads per multiprocessor is 1024. + +A.1.4 + +Specifications for Compute Capability 1.3 +(cid:137) Support for double-precision floating-point numbers. + +A.2 + +Floating-Point Standard + +All compute devices follow the IEEE-754 standard for binary floating-point +arithmetic with the following deviations: + +(cid:137) There is no dynamically configurable rounding mode; however, most of the +operations support IEEE rounding modes, exposed via device functions; + +(cid:137) There is no mechanism for detecting that a floating-point exception has occurred +and all operations behave as if the IEEE-754 exceptions are always masked, and +deliver the masked response as defined by IEEE-754 if there is an exceptional +event; for the same reason, while SNaN encodings are supported, they are not +signaling; + +(cid:137) Absolute value and negation are not compliant with IEEE-754 with respect to + +NaNs; these are passed through unchanged; +(cid:137) For single-precision floating-point numbers only: + +(cid:137) Denormalized numbers are not supported; floating-point arithmetic and + +comparison instructions convert denormalized operands to zero prior to the +floating-point operation; + +(cid:137) Underflowed results are flushed to zero; +(cid:137) The result of an operation involving one or more input NaNs is the quiet + +NaN of bit pattern 0x7fffffff; note that; +(cid:137) Some instructions are not IEEE-compliant: + +NVIDIA OpenCL Programming Guide Version 2.3 + +53 + + + + + + + + Appendix A. Technical Specifications + +(cid:137) Addition and multiplication are often combined into a single multiply-add + +instruction (FMAD), which truncates the intermediate result of the +multiplication; + +(cid:137) Division is implemented via the reciprocal in a non-standard-compliant + +way; + +(cid:137) Square root is implemented via the reciprocal square root in a non- + +standard-compliant way; + +(cid:137) For addition and multiplication, only round-to-nearest-even and + +round-towards-zero are supported via static rounding modes; directed +rounding towards +/- infinity is not supported; + +But, IEEE-compliant software (and therefore slower) implementations are +provided through the following intrinsics from Appendix B: +(cid:137) fma(float, float, float): single-precision fused multiply-add + +with IEEE rounding modes, + +(cid:137) native_recip(float): single-precision reciprocal with IEEE + +rounding modes, + +(cid:137) native_divide(float, float): single-precision division with + +IEEE rounding modes, + +(cid:137) native_sqrt(float): single-precision square root with IEEE + +rounding modes; + +(cid:137) For double-precision floating-point numbers only: + +(cid:137) Round-to-nearest-even is the only supported IEEE rounding mode for + +reciprocal, division, and square root. + +In accordance to the IEEE-754R standard, if one of the input parameters to +fmin() or fmax() is NaN, but not the other, the result is the non-NaN +parameter. + +(cid:137) The conversion of a floating-point value to an integer value in the case where the +floating-point value falls outside the range of the integer format is left undefined +by IEEE-754. For compute devices, the behavior is to clamp to the end of the +supported range. This is unlike the x86 architecture behaves. + +A.3 + +Supported OpenCL Extensions + +All compute devices supports the cl_khr_byte_addressable_store extension. + +Devices of compute capability 1.1 and higher support the +cl_khr_global_int32_base_atomics, cl_khr_global_int32_extended_atomics, +cl_khr_local_int32_base_atomics, and cl_khr_local_int32_extended_atomics +extensions. + +54 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + + + Appendix B. +Mathematical Functions Accuracy + +B.1 + +Standard Functions + +Error bounds in this section are generated from extensive but not exhaustive tests, +so they are not guaranteed bounds. + +B.1.1 + +Single-Precision Floating-Point Functions +Table C-1 lists errors for the standard single-precision floating-point functions. + +The recommended way to round a single-precision floating-point operand to an +integer, with the result being a single-precision floating-point number is rint(), +not round(). The reason is that round() maps to an 8-instruction sequence on +the device, whereas rint() maps to a single instruction. trunc(), ceil(), and +floor() each map to a single instruction as well. + +Table C-1. Mathematical Standard Library Functions with + +Maximum ULP Error + +The maximum error is stated as the absolute value of the difference +in ulps between a correctly rounded single-precision result and the +result returned by the CUDA library function. + +Function +x+y + +x*y + +x/y + +1/x + +1/sqrt(x) +rsqrt(x) + +sqrt(x) + +cbrt(x) + +hypot(x,y) + +exp(x) + +Maximum ulp error +0 (IEEE-754 round-to-nearest-even) +(except when merged into an FMAD) + +0 (IEEE-754 round-to-nearest-even) +(except when merged into an FMAD) + +2 (full range) + +1 (full range) + +2 (full range) + +3 (full range) + +1 (full range) + +3 (full range) + +2 (full range) + + + + + + + + + Appendix B. Mathematical Functions Accuracy + +Function +exp2(x) + +exp10(x) + +expm1(x) + +log(x) + +log2(x) + +log10(x) + +log1p(x) + +sin(x) + +cos(x) + +tan(x) + +sincos(x,cptr) + +asin(x) + +acos(x) + +atan(x) + +atan2(y,x) + +sinh(x) + +cosh(x) + +tanh(x) + +asinh(x) + +acosh(x) + +atanh(x) + +pow(x,y) + +erf(x) + +erfc(x) + +erfinv(x) + +erfcinv(x) + +lgamma(x) + +tgamma(x) + +fma(x,y,z) + +frexp(x,exp) + +ldexp(x,exp) + +scalbn(x,n) + +scalbln(x,l) + +logb(x) + +ilogb(x) + +fmod(x,y) + +remainder(x,y) + +remquo(x,y,iptr) + +modf(x,iptr) + +fdim(x,y) + +trunc(x) + +round(x) + +Maximum ulp error +2 (full range) + +2 (full range) + +1 (full range) + +1 (full range) + +3 (full range) + +3 (full range) + +2 (full range) + +2 (full range) + +2 (full range) + +4 (full range) + +2 (full range) + +4 (full range) + +3 (full range) + +2 (full range) + +3 (full range) + +3 (full range) + +2 (full range) + +2 (full range) + +3 (full range) + +4 (full range) + +3 (full range) + +8 (full range) + +3 (full range) + +8 (full range) + +5 (full range) + +7 (full range) + +6 (outside interval -10.001 ... -2.264; larger inside) + +11 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +56 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + Appendix B. Mathematical Functions Accuracy + +Function +rint(x) + +nearbyint(x) + +ceil(x) + +floor(x) + +lrint(x) + +lround(x) + +llrint(x) + +llround(x) + +Maximum ulp error +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +B.1.2 + +Double-Precision Floating-Point Functions +Table C-2 lists errors for the standard double-precision floating-point functions. + +These errors only apply when compiling for devices with native double-precision +support. When compiling for devices without such support, such as devices of +compute capability 1.2 and lower, the double type gets demoted to float by +default and the double-precision math functions are mapped to their single- +precision equivalents. + +The recommended way to round a double-precision floating-point operand to an +integer, with the result being a double-precision floating-point number is rint(), +not round(). The reason is that round() maps to an 8-instruction sequence on +the device, whereas rint() maps to a single instruction. trunc(), ceil(), and +floor() each map to a single instruction as well. + +Table C-2. Mathematical Standard Library Functions with + +Maximum ULP Error + +The maximum error is stated as the absolute value of the difference +in ulps between a correctly rounded double-precision result and the +result returned by the CUDA library function. + +Function +x+y + +x*y + +x/y + +1/x + +sqrt(x) + +rsqrt(x) + +cbrt(x) + +hypot(x,y) + +exp(x) + +exp2(x) + +exp10(x) + +expm1(x) + +log(x) + +log2(x) + +Maximum ulp error +0 (IEEE-754 round-to-nearest-even) + +0 (IEEE-754 round-to-nearest-even) + +0 (IEEE-754 round-to-nearest-even) + +0 (IEEE-754 round-to-nearest-even) + +0 (IEEE-754 round-to-nearest-even) + +1 (full range) + +1 (full range) + +2 (full range) + +1 (full range) + +1 (full range) + +1 (full range) + +1 (full range) + +1 (full range) + +1 (full range) + +NVIDIA OpenCL Programming Guide Version 2.3 + +57 + + + + + + + + Appendix B. Mathematical Functions Accuracy + +Function +log10(x) + +log1p(x) + +sin(x) + +cos(x) + +tan(x) + +sincos(x,sptr,cptr) + +asin(x) + +acos(x) + +atan(x) + +atan2(y,x) + +sinh(x) + +cosh(x) + +tanh(x) + +asinh(x) + +acosh(x) + +atanh(x) + +pow(x,y) + +erf(x) + +erfc(x) + +erfinv(x) + +erfcinv(x) + +lgamma(x) + +tgamma(x) + +fma(x,y,z) + +frexp(x,exp) + +ldexp(x,exp) + +scalbn(x,n) + +scalbln(x,l) + +logb(x) + +ilogb(x) + +fmod(x,y) + +remainder(x,y) + +remquo(x,y,iptr) + +modf(x,iptr) + +fdim(x,y) + +trunc(x) + +round(x) + +rint(x) + +nearbyint(x) + +ceil(x) + +floor(x) + +lrint(x) + +Maximum ulp error +1 (full range) + +1 (full range) + +2 (full range) + +2 (full range) + +2 (full range) + +2 (full range) + +2 (full range) + +2 (full range) + +2 (full range) + +2 (full range) + +1 (full range) + +1 (full range) + +1 (full range) + +2 (full range) + +2 (full range) + +2 (full range) + +2 (full range) + +2 (full range) + +7 (full range) + +8 (full range) + +8 (full range) + +4 (outside interval -11.0001 ... -2.2637; larger inside) + +8 (full range) + +0 (IEEE-754 round-to-nearest-even) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +0 (full range) + +58 + +NVIDIA OpenCL Programming Guide Version 2.3 + + + + + + Appendix B. Mathematical Functions Accuracy + +Function +lround(x) + +llrint(x) + +llround(x) + +Maximum ulp error +0 (full range) + +0 (full range) + +0 (full range) + +B.2 + +Native Functions + +Table C-3 lists the native single-precision floating-point functions supported on the +CUDA architecture. + +Both the regular floating-point division and native_divide(x,y) have the same +accuracy, but for 2126 < y < 2128, native_divide(x,y) delivers a result of zero, +whereas the regular division delivers the correct result to within the accuracy stated +in Table C-3. Also, for 2126 < y < 2128, if x is infinity, native_divide(x,y) +delivers a NaN (as a result of multiplying infinity by zero), while the regular division +returns infinity. + +Table C-3. Single-Precision Floating-Point Native Functions + +with Respective Error Bounds + +Function +native_recip(x) + +native_sqrt(x) + +Error bounds +IEEE-compliant. + +IEEE-compliant. + +native_divide(x,y) + +For y in [2-126, 2126], the maximum ulp error is 2. + +native_exp(x) + +native_exp10(x) + +native_log(x) + +native_log2(x) + +native_log10(x) + +native_sin(x) + +native_cos(x) + +native_tan(x) + +native_pow(x,y) + +The maximum ulp error is +2 + floor(abs(1.16 * x)). + +The maximum ulp error is +2 + floor(abs(2.95 * x)). +For x in [0.5, 2], the maximum absolute error is 2- +21.41, otherwise, the maximum ulp error is 3. +For x in [0.5, 2], the maximum absolute error is 2-22, +otherwise, the maximum ulp error is 2. +For x in [0.5, 2], the maximum absolute error is 2-24, +otherwise, the maximum ulp error is 3. + +For x in [-π, π], the maximum absolute error is 2-21.41, +and larger otherwise. +For x in [-π, π], the maximum absolute error is 2-21.19, +and larger otherwise. + +Derived from its implementation as +native_sin(x) * (1 / native_cos(x)). + +Derived from its implementation as +exp2(y * native_log2(x)). + +NVIDIA OpenCL Programming Guide Version 2.3 + +59 + + + + + + + + + + Notice + +ALL NVIDIA DESIGN SPECIFICATIONS, REFERENCE BOARDS, FILES, DRAWINGS, DIAGNOSTICS, LISTS, AND +OTHER DOCUMENTS (TOGETHER AND SEPARATELY, “MATERIALS”) ARE BEING PROVIDED “AS IS.” NVIDIA +MAKES NO WARRANTIES, EXPRESSED, IMPLIED, STATUTORY, OR OTHERWISE WITH RESPECT TO THE +MATERIALS, AND EXPRESSLY DISCLAIMS ALL IMPLIED WARRANTIES OF NONINFRINGEMENT, +MERCHANTABILITY, AND FITNESS FOR A PARTICULAR PURPOSE. + +Information furnished is believed to be accurate and reliable. However, NVIDIA Corporation assumes no +responsibility for the consequences of use of such information or for any infringement of patents or other +rights of third parties that may result from its use. No license is granted by implication or otherwise under any +patent or patent rights of NVIDIA Corporation. Specifications mentioned in this publication are subject to +change without notice. This publication supersedes and replaces all information previously supplied. NVIDIA +Corporation products are not authorized for use as critical components in life support devices or systems +without express written approval of NVIDIA Corporation. + +Trademarks + +NVIDIA, the NVIDIA logo, GeForce, Tesla, and Quadro are trademarks or registered trademarks of NVIDIA +Corporation. Other company and product names may be trademarks of the respective companies with which +they are associated. + +Copyright + +© 2007-2008 NVIDIA Corporation. All rights reserved. + +This work incorporates portions of on an earlier work: Scalable Parallel Programming with CUDA, in ACM +Queue, VOL 6, No. 2 (March/April 2008), © ACM, 2008. http://mags.acm.org/queue/20080304/?u1=texterity" + +NVIDIA Corporation +2701 San Tomas Expressway +Santa Clara, CA 95050 +www.nvidia.com + + + + \ No newline at end of file