GPU Computing: Introduction - KIT

3y ago
24 Views
2 Downloads
1.07 MB
12 Pages
Last View : 13d ago
Last Download : 3m ago
Upload by : Arnav Humphrey
Transcription

GPU Computing: IntroductionDipl.-Ing. Jan Novák Dipl.-Inf. Gábor Liktor†Prof. Dr.-Ing. Carsten Dachsbacher‡AbstractExploiting the vast horse power of contemporary GPUs for general purpose applications has become a must for any real time orinteractive application nowadays. Current computer games use theGPUs not only for rendering graphics, but also for collision detection, physics, or artificial intelligence. General purpose computing on GPUs (GPGPU) has also penetrated the field of scientificcomputing enabling real time experience of large scale fluid simulations, medical visualization, signal processing, etc. This lectureintroduces the concepts of programming graphics cards for nongraphical applications, such as data sorting, image filtering (e.g.denoising, sharpening), or physically based simulations.In this very first assignment you will briefly hear about the history of GPU computing and the motivations that drive us to harnesscontemporary GPUs for general purpose computation. We will introduce the architectural considerations and constraints of contemporary GPUs that have to be reflected in the algorithm, shall it beefficient. You will also learn the basic concepts of the OpenCL programming language wrapped in a simple framework that we willbe using during the course. All the knowledge you gain by readingthis paper will be then applied in two simple starting assignments.1Parallel ProgrammingIn few cases, the transition between sequential and parallel environment can be trivial. Consider for example a simple particle simulation, where particles are affected only by a gravity field. Sequentialalgorithm would iterate over all particles at each time step and perform some kind of integration technique to compute the new position (and velocity) of each particle. On a parallel architecture, wecan achieve the same by creating a number of threads, each handling exactly one particle. If the parallel hardware contains enoughof processing units, all particles can be processed in a single parallelstep speeding up the simulation by a factor of N .Unfortunately, not all problems can be parallelized in such a trivialmanner; even worse, the number of such problems is quite low. Inpractical applications, we are facing much harder tasks that often require adjusting the algorithm or even reformulating of the problem.Even though each problem can be essentially unique, there is a setof well working paradigms that significantly increase the chancesof successfully parallelizing an algorithm without too much of frustration and a risk of being fired. The road to victory consists of thefollowing steps:1. Decompose the problem into a set of smaller tasks and identify whether each of them can be (easily) parallelized or not.Whereas some parts can always be identified as embarrassingly parallel, others may require serialization of processingunits and/or inter-thread communication. Finding the rightgranularity already at the beginning can save a lot of effort inthe later development.2. Estimate the trade-offs of parallelizing the inherently serialparts of the algorithm. If such parts can be efficiently processed on the CPU and the transfer between CPU and GPU isonly marginal (compared to the rest of the algorithm), there e-mail:jan.novak@kit.edugabor.liktor@kit.edu‡ e-mail: dachsbacher@kit.edu† e-mail:is no need to parallelize these parts at any cost, as it may result in overall slow-down. On the other hand, even a slightlyslower GPU implementation can be preferable if transferringdata between RAM and GPU memory is costly.3. Reformulate the problem if the solution does not seem tofit the architecture well. This is of course not possible in allcases, but using a different data layout, order of access, orsome preprocessing can tackle the problem more efficiently.An example from the early GPGPU: architectures at the timedid not allow for efficient scatter operations, the key to successat that time was to use a gather operation instead: collectingthe data from neighbors instead of distributing it to them.4. Pick the right algorithm for your application. Given a problem we can typically find several algorithms accomplishingthe same using different approaches. It is necessary to compare them in terms of storage and bandwidth requirements,arithmetic intensity, and cost and step efficiency.5. Profile and analyze your implementation. There are oftenseveral options how we can optimize an initial implementation leading to significant speed up. Make sure that the accesses to the global device memory are aligned, kernels do notwaste registers, the transfer between CPU and GPU is minimized, and the number of threads enables high occupancy.These are only some basic concepts that we will (and someothers) introduce during individual assignments. When optimizing you should always obey Amdahl’s Law and focus onparts that consume most of the execution time, rather thanthose that can be easily optimized.Parallel computing is a comprehensive and complex area of computer science that is hard to master without practice and experience.During this course we will try teach you a little bit of computationalthinking through a number of assignments with emphasis on theright choice of algorithms and optimizations. Nevertheless, as thescope of the assignments must be limited (unfortunately), we pointyou to some useful literature and on-line seminars that can serve assupplementary material: Programming Massively Parallel Processors, A Hands-on Approach, David B. Kirk, Wen-mei W. Hwu., Morgan Kaufmann, 2010 NVIDIA OpenCL Programming ng.html GPU Computing Online Seminarshttp://developer.nvidia.com/object/gpu computing online.html1.1History of Graphics AcceleratorsHistorical beginnings of modern graphics processing units dateback to mid eighties, when the Amiga Corporation released theirfirst computer featuring a device that would be nowadays recognized as a full graphics accelerator. Prior to this turning point, allthe computers generated the graphics content on central processingunit (CPU). Offloading the computation of graphics to a dedicateddevice allowed higher specialization of the hardware and relievedthe computational requirements on the CPU. By 1995, replaceablegraphics cards with fixed-function accelerators surpassed expensivegeneral-purpose coprocessors, which have completely faded awayfrom the market in next few years (note the historical trend that wascompletely opposite to contemporary evolution of GPUs).

Large number of manufacturers and increasing demand onhardware-accelerated 3D graphics led to an establishment of twoapplication programming interface (API) standards named OpenGLand DirectX. Whereas the first did not restrict its usage to a particular hardware and benefited from cutting-edge technologies of individual card series, the latter was usually one step behind due toits strict marketing policy targeting only a subset of vendors. Nevertheless, the difference quickly disappeared as Microsoft startedworking closely with GPU developers reaching a widespread adoption of its DirectX 5.0 in gaming market.After 2001, GPU manufacturers enhanced the accelerators byadding support for programmable shading, allowing game developers and designers to adjust rendering algorithms to produce customized results. GPUs were equipped with conditional statements,loops, unordered accesses to the memory (gather and later scatteroperations), and became moderately programmable devices. Suchfeatures enabled first attempts to exploit the graphics dedicatedhardware for computing non-graphical tasks. The true revolutionin the design of graphics cards came in 2007, when both marketleading vendors, NVIDIA and ATI, dismissed the idea of separatespecialized (vertex and fragment) shaders and replaced them witha single set of unified processing units. Instead of processing vertices and fragments at different units, the computation is nowadaysperformed on one set of unified processors only. Furthermore, thesimplified architecture allows less complicated hardware design,which can be manufactured with shorter and faster silicon technology. Rendering of graphics is carried out with respect to the traditional graphics pipeline, where the GPU consecutively utilizes theset of processing units for vertex operations, geometry processing,fragment shading, and possibly some others. Thanks to the unifieddesign, porting general tasks is nowadays less restrictive. Note thehistorical back-evolution: though highly advanced, powerful, andmuch more mature, modern GPUs are in some sense conceptuallyvery similar to graphics accelerators manufactured before 1995.1.2Programming Languages and EnvironmentsIn order to write a GPU program, we first need to choose a suitable programming language. Besides others, there are three mainstream shading languages (GLSL, HLSL, and Cg) enabling general computing via mapping the algorithm to the traditional graphics pipeline. Since the primary target is processing of graphics,aka shading, programs written in these languages tightly follow thegraphics pipeline and require the programmer to handle the data asvertices and fragments and store resources and results in buffers andtextures. To hide the architecture of the underlying hardware, various research groups created languages for general computation onGPUs and multi-core CPUs. Among the most popular belong theSh, Brook, and RapidMind, which yielded success mostly in otherareas than computer graphics. Nevertheless, none of them brought amajor breakthrough, mostly due to the inherent limitations imposedfrom hardware restrictions.The situation improved with the unified architecture of GPUs. Contemporary NVIDIA graphics cards automatically support CUDAprogramming language, and the platform-independent OpenCLcloses the gap for the remaining vendors, supporting even heterogenous computation on multiple central and graphics processing unitsat the same time. Unlike the shading languages, both CUDA andOpenCL enable true general purpose computation without resortingto the traditional graphics pipeline. The syntax and semantic rulesare inherited from C with a few additional keywords to specify different types of execution units and memories.Figure 1: Architecture of the GF100 (Fermi) streaming multiprocessor. Image courtesy of NVIDIA.2Computing Architecture of Modern GPUsBefore we introduce the OpenCL programming language that willbe used throughout this course, we will outline the architecture ofmodern GPUs. You should then clearly understand how the abstract language maps to the actual hardware. As most of the computers in the lab are equipped with NVIDIA graphics cards, andalso because the NVIDIA’s Compute Unified Device Architecture(CUDA) is more open to general computing, we will describe theindividual parts of the compute architecture in the context of contemporary NVIDIA GPUs.2.1Streaming DesignModern many-core GPUs (those developed by ATI and NVIDIA,not Intel’s Larrabee, which was supposed to be conceptually different) consist of a several streaming multiprocessors (SM) that operate on large sets of streaming data. Each multiprocessor contains a number of streaming processors (SP). To perform the actualcomputation, individual SPs are equipped with several arithmethic(ALU), and floating point (FPU) units. The streaming multiprocessor is further provided with several load, store, and special functionunits, which are used for loading and storing data, and transcendental functions (e.g. sine, cosine, etc.) respectively. In order toexecute the code, each SM uses an instruction cache, from whichthe warp scheduler and dispatch units fetch instructions and matchthem with GPU threads to be executed on the SM. Data can bestored either in registers, or in one of the very fast on-chip memories, depending on the access and privacy restrictions. As thecapacity of the on-chip memory is highly limited, contemporaryGPUs have gigabytes of device memory. We will provide somemore detail about different storage options in Section 2.2. Figure 1illustrates the architecture of a single streaming multiprocessor.

Constant Memory represents a specific part of the devicememory, which allows to store limited amount (64 KB) ofconstant data (on CUDA called symbols). Similarly to thetexture memory, the accesses are cached but only reading isallowed. Constant memory should be used for small variablesthat are shared among all threads and do no require interpolation.DeviceMultiprocessor NMultiprocessor 2Multiprocessor 1Shared ocessor 1Processor 2 Processor MConstantCacheTextureCacheDevice MemoryFigure 2: Memory hierarchy of CUDA GPUs. Image courtesyof NVIDIA.2.2Memory ModelMemory model of contemporary NVIDIA graphics cards is shownon Figure 2. Different memory spaces can be classified regardingtheir degree of privacy. Each thread has a private local memory thatcannot be shared. For cooperation of threads within a block sharedmemory can be used. Finally, an arbitrary exchange of data betweenall threads can only be achieved via a transfer through the globalmemory. Since different memory spaces have different parameters,such as latency and capacity, we provide a brief description of eachin the following sections.2.2.1Device MemoryThe most prominent feature of the device memory is its high capacity, which in case of the newest GPUs reaches up to 4 GB. Onthe other hand, all memory spaces reserved in the device memoryexhibit very high latency (400 to 600 clock cycles) prohibiting anextensive usage, when high performance is requested. Individualspaces are listed below. Global Memory is the most general space allowing bothreading and writing data. Accesses to the global memorywere prior to Fermi GPUs not cached. Despite the automaticcaching we should try to use well-defined addressing to coalesce the accesses into a single transaction and minimize theoverall latency. Texture Memory, as its name suggests, is optimized for storing textures. This type of storage is a read-only memory capable of automatically performing bilinear and trilinear interpolation of neighboring values (when floating point coordinatesare used for addressing). Data fetches from the memory arecached, efficiently hiding the latency when multiple threadsaccess the same item. Local Memory space is automatically allocated during theexecution of kernels to provide the threads with storage forlocal variables that do not fit into the registers. Since localmemory is not cached, the accesses are as expensive as accessing the global memory; however, the latency is partiallyhidden by automatic coalescing.All previously mentioned device spaces, except for the local memory, are allocated and initialized by the host. Threads can onlyoutput the results of computation into the global memory; hence,it is used for exchanging data between successive kernels. Texture memory should be used for read-only data with spatial locality,whereas constant memory is suitable for common parameters andstatic variables.2.2.2On-chip MemoryThe counterpart of the device memory is the on-chip memory,which manifests very low latency. Since it is placed directly onthe multiprocessor, its capacity is very low allowing only a limitedand specific usage, mostly caching and fast inter-thread communication. Registers are one of the most important features of the GPUwhen it comes to complex algorithms. If your program requires to many registers, it will hurt the performance since thewarp scheduler cannot schedule enough threads on the SM.Multiprocessors on contemporary CUDA GPUs are equippedwith 16384 (32768 on Fermi) registers with zero latency. Shared Memory, sometimes also called parallel data cache,or group-shared memory (in the context of DirectX), or local memory (in OpenCL), serves as a low latency storage forcooperation between threads. Its capacity of 16 KB (up to48 KB on Fermi) is split between all blocks running on themultiprocessor in pseudo-parallel. The memory is composedof 16 (32 on Fermi) banks that can be accessed simultaneously; therefore, the threads must coordinate its accesses toavoid conflicts and subsequent serialization. The lifetime ofvariables in shared memory equal to the lifetime of the block,so any variables left in the memory after the block has beenprocessed are automatically discarded. Texture Cache hides the latency of accessing the texturememory. The cache is optimized for 2D spatial locality, sothe highest performance is achieved when threads of the samewarp access neighboring addresses. The capacity of the cachevaries between 6 and 8 KB per multiprocessor, depending onthe graphics card. Constant Cache is similar to texture cache: it caches the dataread from the constant memory. The cache is shared by allprocessing units within the SM and its capacity on CUDAcards is 8 KB.The only on-chip memory available to the programmer is the sharedmemory. Usage of both caches and registers is managed automatically by the memory manager, hiding any implementation detailsfrom the programmer.

3The OpenCL PlatformProgrammers have often been challenged by the task of solving thesame problem on different architectures. Classical language standards, like ANSI C or C , made life a lot easier: instead of using assembly instructions, the same high-level code could be compiled to any specific CPU ISA (Instruction Set Architecture). Asthe hardware generations evolved and took different directions, thegoal to have ”one code to rule them all” became more and more difficult to reach. The growth of clock rates of CPUs is slowing down,so the only way to continue the trend of Moore’s law remained toincrease the number of cores on the chip, thus making the executionparallel. A classical single-threaded program uses only small partof the available resources, forcing programmers to adopt new algorithms from the world of distributed systems. Contemporary GPUsare offering an efficient alternative for wide range of programmingproblems via languages very similar to C and C . As the different platforms have different features for optimization, (out-of-orderexecution, specialized memory for textures), the programmer needsmore and more specific knowledge about the hardware than beforefor low-level optimization.Open Computing Language (OpenCL) is a new standard in parallelcomputing that targets simultaneous computation on heterogeneousplatforms. It has been proposed by Apple Inc. and developed injoint cooperation with other leading companies in the field (Intel,NVIDIA, AMD, IBM, Motorola, and many others). Since it is anopen standard (from 2008 maintained by the Khronos Group, whichalso takes care of OpenGL and OpenAL) it promises cross-platformapplicability and support of many hardware vendors. By employing abstraction layers, an OpenCL application can be mapped tovarious hardware and can take different execution paths based onthe available device capabilities. The programmer should still behighly familiar with parallelization, but can exploit the features ofthe underlying architecture by only knowing the fact that it implements some parts of a standardized model. From now on, we shallexclusively focus on GPGPU programming using OpenCL, but weshould always keep in mind that using the same abstraction, unified parallelization is possible for heterogeneous, multi-CPU-GPUsystems as well.Even though OpenCL strives to provide an abstract foundation forgeneral purpose computing, it still requires the programmer to follow a set of paradigms, arising from the common features of variousarchitectures. These are described within the context of four abstract models (Platform, Execution, Memory, and Programmingmodels) that OpenCL uses to hide hardware complexity.3.1Platform Model - Host and DevicesIn the context of parallel programming, we need to distinguish between the hardware that performs the actual computation (device)and th

GPU Computing: Introduction Dipl.-Ing. Jan Nov ak Dipl.-Inf. Gabor Liktor y Prof. Dr.-Ing. Carsten Dachsbacherz Abstract Exploiting the vast horse power of contemporary GPUs for gen-eral purpose applications has become a must for any real time or interactive application nowadays. Current computer games use the

Related Documents:

2 Valve body KIT M100201 KIT M100204 KIT M100211 KIT M100211 KIT M100218 KIT M300222 7 Intermediate cover (double diaphragm) - - - KIT M110098 KIT M110100 KIT M110101 4 Top cover KIT M110082 KIT M110086 KIT M110092 KIT M110082 KIT M110082 KIT M110082 5 Diaphragm KIT DB 16/G KIT DB 18/G KIT DB 112/G - - - 5 Viton Diaphragm KIT DB 16V/S KIT

the gpu computing era gpu computing is at a tipping point, becoming more widely used in demanding consumer applications and high-performance computing.this article describes the rapid evolution of gpu architectures—from graphics processors to massively parallel many-core multiprocessors, recent developments in gpu computing architectures, and how the enthusiastic

GPU Tutorial 1: Introduction to GPU Computing Summary This tutorial introduces the concept of GPU computation. CUDA is employed as a framework for this, but the principles map to any vendor’s hardware. We provide an overview of GPU computation, its origins and development, before presenting both the CUDA hardware and software APIs. New Concepts

OpenCV GPU header file Upload image from CPU to GPU memory Allocate a temp output image on the GPU Process images on the GPU Process images on the GPU Download image from GPU to CPU mem OpenCV CUDA example #include opencv2/opencv.hpp #include <

Introduction to GPU computing Felipe A. Cruz Nagasaki Advanced Computing Center Nagasaki University, Japan. Felipe A. Cruz Nagasaki University The GPU evolution The Graphic Processing Unit (GPU) is a processor that was specialized for processing graphics. The GPU has recently evolved towards a more flexible architecture.

GPU Computing in Matlab u Included in the Parallel Computing Toolbox. u Extremely easy to use. To create a variable that can be processed using the GPU, use the gpuArray function. u This function transfers the storage location of the argument to the GPU. Any functions which use this argument will then be computed by the GPU.

limitation, GPU implementers made the pixel processor in the GPU programmable (via small programs called shaders). Over time, to handle increasing shader complexity, the GPU processing elements were redesigned to support more generalized mathematical, logic and flow control operations. Enabling GPU Computing: Introduction to OpenCL

Will Landau (Iowa State University) Introduction to GPU computing for statisticicans September 16, 2013 20 / 32. Introduction to GPU computing for statisticicans Will Landau GPUs, parallelism, and why we care CUDA and our CUDA systems GPU computing with R CUDA and our CUDA systems Logging in