Skip to content

Latest commit

 

History

History
469 lines (380 loc) · 25.2 KB

OffloadingDesign.rst

File metadata and controls

469 lines (380 loc) · 25.2 KB

Offloading Design & Internals

This document describes the Clang driver and code generation steps for creating offloading applications. Clang supports offloading to various architectures using programming models like CUDA, HIP, and OpenMP. The purpose of this document is to illustrate the steps necessary to create an offloading application using Clang.

Note

This documentation describes Clang's behavior using the new offloading driver which. This currently must be enabled manually using -fopenmp-new-driver.

Clang supports OpenMP target offloading to several different architectures such as NVPTX, AMDGPU, X86_64, Arm, and PowerPC. Offloading code is generated by Clang and then executed using the libomptarget runtime and the associated plugin for the target architecture, e.g. libomptarget.rtl.cuda. This section describes the steps necessary to create a functioning device image that can be loaded by the OpenMP runtime. More information on the OpenMP runtimes can be found at the OpenMP documentation page.

The goal of offloading compilation is to create an executable device image that can be run on the target device. OpenMP offloading creates executable images by compiling the input file for both the host and the target device. The output from the device phase then needs to be embedded into the host to create a fat object. A special tool then needs to extract the device code from the fat objects, run the device linking step, and embed the final image in a symbol the host can use to register the library and access the symbols on the device.

The compiler performs the following high-level actions to generate offloading code:

The first step in compilation is to generate offloading entries for the host. This information is used to identify function kernels or global values that will be provided by the device. Blocks contained in a #pragma omp target or symbols inside a #pragma omp declare target directive will have offloading entries generated. The following table shows the :ref:`offload entry structure <table-tgt_offload_entry_structure>`.

__tgt_offload_entry Structure
Type Identifier Description
void* addr Address of global symbol within device image (function or global)
char* name Name of the symbol
size_t size Size of the entry info (0 if it is a function)
int32_t flags Flags associated with the entry (see :ref:`table-offload_entry_flags`)
int32_t reserved Reserved, to be used by the runtime library.

The address of the global symbol will be set to the appropriate value by the runtime once the device image is loaded. The flags are set to indicate the handling required for the offloading entry. If the offloading entry is an entry to a target region it can have one of the following :ref:`entry flags <table-offload_entry_flags>`.

Target Region Entry Flags
Name Value Description
OMPTargetRegionEntryTargetRegion 0x00 Mark the entry as generic target region
OMPTargetRegionEntryCtor 0x02 Mark the entry as a global constructor
OMPTargetRegionEntryDtor 0x04 Mark the entry as a global destructor

If the offloading entry is a global variable, indicated by a non-zero size, it will instead have one of the following :ref:`global <table-offload_global_flags>` flags.

Target Region Global
Name Value Description
OMPTargetGlobalVarEntryTo 0x00 Mark the entry as a 'to' attribute (w.r.t. the to clause)
OMPTargetGlobalVarEntryLink 0x01 Mark the entry as a 'link' attribute (w.r.t. the link clause)

The target offload entries are used by the runtime to access the device kernels and globals that will be provided by the final device image. Each offloading entry is set to use the omp_offloading_entries section. When the final application is created the linker will provide the __start_omp_offloading_entries and __stop_omp_offloading_entries symbols which are used to create the :ref:`final image <Device Binary Wrapping>`.

This information is by the device compilation stage to determine which symbols need to be exported from the device. We use the omp_offload.info metadata node to pass this information device compilation stage.

Accessing the entries in the device is done using the address field in the :ref:`offload entry<table-tgt_offload_entry_structure>`. The runtime will set the address to the pointer associated with the device image during runtime initialization. This is used to call the corresponding kernel function when entering a #pragma omp target region. For variables, the runtime maintains a table mapping host pointers to device pointers. Global variables inside a #pragma omp target reclare directive are first initialized to the host's address. Once the device address is initialized we insert it into the table to map the host address to the device address.

We generate structures to hold debugging information that is passed to libomptarget. This allows the front-end to generate information the runtime library uses for more informative error messages. This is done using the standard :ref:`identifier structure <table-ident_t_structure>` used in libomp and libomptarget. This is used to pass information and source locations to the runtime.

ident_t Structure
Type Identifier Description
int32_t reserved Reserved, to be used by the runtime library.
int32_t flags Flags used to indicate some features, mostly unused.
int32_t reserved Reserved, to be used by the runtime library.
int32_t reserved Reserved, to be used by the runtime library.
char* psource Program source information, stored as ";filename;function;line;column;;\0"

If debugging information is enabled, we will also create strings to indicate the names and declarations of variables mapped in target regions. These have the same format as the source location in the :ref:`identifier structure <table-ident_t_structure>`, but the filename is replaced with the variable name.

The input file is compiled for each active device toolchain. The device compilation stage is performed differently from the host stage. Namely, we do not generate any offloading entries. This is set by passing the -fopenmp-is-device flag to the front-end. We use the host bitcode to determine which symbols to export from the device. The bitcode file is passed in from the previous stage using the -fopenmp-host-ir-file-path flag. Compilation is otherwise performed as it would be for any other target triple.

When compiling for the OpenMP device, we set the visibility of all device symbols to be protected by default. This improves performance and prevents a class of errors where a symbol in the target device could preempt a host library.

The OpenMP runtime library is linked in during compilation to provide the implementations for standard OpenMP functionality. For GPU targets this is done by linking in a special bitcode library during compilation, (e.g. libomptarget-nvptx64-sm_70.bc) using the -mlink-builtin-bitcode flag. Other device libraries, such as CUDA's libdevice, are also linked this way. If the target is a standard architecture with an existing libomp implementation, that will be linked instead. Finally, device tools are used to create a relocatable device object file that can be embedded in the host.

A fat binary is a binary file that contains information intended for another device. We create a fat object by embedding the output of the device compilation stage into the host as a named section. The output from the device compilation is passed to the host backend using the -fembed-offload-object flag. This inserts the object as a global in the host's IR. The section name contains the target triple and architecture that the data corresponds to for later use. Typically we will also add an extra string to the section name to prevent it from being merged with other sections if the user performs relocatable linking on the object.

@llvm.embedded.object = private constant [1 x i8] c"\00", section ".llvm.offloading.nvptx64.sm_70."

The device code will then be placed in the corresponding section one the backend is run on the host, creating a fat object. Using fat objects allows us to treat offloading objects as standard host objects. The final object file should contain the following :ref:`offloading sections <table-offloading_sections>`. We will use this information when :ref:`Device Linking`.

Offloading Sections
Section Description
omp_offloading_entries Offloading entry information (see :ref:`table-tgt_offload_entry`)
.llvm.offloading.<triple>.<arch> Embedded device object file for the target device and architecture

Objects containing :ref:`table-offloading_sections` require special handling to create an executable device image. This is done using a Clang tool, see :doc:`ClangLinkerWrapper` for more information. This tool works as a wrapper over the host linking job. It scans the input object files for the offloading sections and runs the appropriate device linking action. The linked device image is then :ref:`wrapped <Device Binary Wrapping>` to create the symbols used to load the device image and link it with the host.

The linker wrapper tool supports linking bitcode files through link time optimization (LTO). This is used whenever the object files embedded in the host contain LLVM bitcode. Bitcode will be embedded for architectures that do not support a relocatable object format, such as AMDGPU or SPIR-V, or if the user passed in -foffload-lto.

Various structures and functions are used to create the information necessary to offload code on the device. We use the :ref:`linked device executable <Device Linking>` with the corresponding offloading entries to create the symbols necessary to load and execute the device image.

Several different structures are used to store offloading information. The :ref:`device image structure <table-device_image_structure>` stores a single linked device image and its associated offloading entries. The offloading entries are stored using the __start_omp_offloading_entries and __stop_omp_offloading_entries symbols generated by the linker using the :ref:`table-tgt_offload_entry`.

__tgt_device_image Structure
Type Identifier Description
void* ImageStart Pointer to the target code start
void* ImageEnd Pointer to the target code end
__tgt_offload_entry* EntriesBegin Begin of table with all target entries
__tgt_offload_entry* EntriesEnd End of table (non inclusive)

The target :ref:`target binary descriptor <table-target_binary_descriptor>` is used to store all binary images and offloading entries in an array.

__tgt_bin_desc Structure
Type Identifier Description
int32_t NumDeviceImages Number of device types supported
__tgt_device_image* DeviceImages Array of device images (1 per dev. type)
__tgt_offload_entry* HostEntriesBegin Begin of table with all host entries
__tgt_offload_entry* HostEntriesEnd End of table (non inclusive)

:ref:`table-global_variables` lists various global variables, along with their type and their explicit ELF sections, which are used to store device images and related symbols.

Global Variables
Variable Type ELF Section Description
__start_omp_offloading_entries __tgt_offload_entry .omp_offloading_entries Begin symbol for the offload entries table.
__stop_omp_offloading_entries __tgt_offload_entry .omp_offloading_entries End symbol for the offload entries table.
__dummy.omp_offloading.entry __tgt_offload_entry .omp_offloading_entries Dummy zero-sized object in the offload entries section to force linker to define begin/end symbols defined above.
.omp_offloading.device_image __tgt_device_image .omp_offloading_entries ELF device code object of the first image.
.omp_offloading.device_image.N __tgt_device_image .omp_offloading_entries ELF device code object of the (N+1)th image.
.omp_offloading.device_images __tgt_device_image .omp_offloading_entries Array of images.
.omp_offloading.descriptor __tgt_bin_desc .omp_offloading_entries Binary descriptor object (see :ref:`binary_descriptor`)

This object is passed to the offloading runtime at program startup and it describes all device images available in the executable or shared library. It is defined as follows:

__attribute__((visibility("hidden")))
extern __tgt_offload_entry *__start_omp_offloading_entries;
__attribute__((visibility("hidden")))
extern __tgt_offload_entry *__stop_omp_offloading_entries;
static const char Image0[] = { <Bufs.front() contents> };
...
static const char ImageN[] = { <Bufs.back() contents> };
static const __tgt_device_image Images[] = {
  {
    Image0,                            /*ImageStart*/
    Image0 + sizeof(Image0),           /*ImageEnd*/
    __start_omp_offloading_entries,    /*EntriesBegin*/
    __stop_omp_offloading_entries      /*EntriesEnd*/
  },
  ...
  {
    ImageN,                            /*ImageStart*/
    ImageN + sizeof(ImageN),           /*ImageEnd*/
    __start_omp_offloading_entries,    /*EntriesBegin*/
    __stop_omp_offloading_entries      /*EntriesEnd*/
  }
};
static const __tgt_bin_desc BinDesc = {
  sizeof(Images) / sizeof(Images[0]),  /*NumDeviceImages*/
  Images,                              /*DeviceImages*/
  __start_omp_offloading_entries,      /*HostEntriesBegin*/
  __stop_omp_offloading_entries        /*HostEntriesEnd*/
};

The global constructor (.omp_offloading.descriptor_reg()) registers the device images with the runtime by calling the __tgt_register_lib() runtime function. The constructor is explicitly defined in .text.startup section and is run once when the program starts. Similarly, the global destructor (.omp_offloading.descriptor_unreg()) calls __tgt_unregister_lib() for the destructor and is also defined in .text.startup section and run when the program exits.

This section contains a simple example of generating offloading code using OpenMP offloading. We will use a simple ZAXPY BLAS routine.

#include <complex>

using complex = std::complex<double>;

void zaxpy(complex *X, complex *Y, complex D, std::size_t N) {
#pragma omp target teams distribute parallel for
  for (std::size_t i = 0; i < N; ++i)
    Y[i] = D * X[i] + Y[i];
}

int main() {
  const std::size_t N = 1024;
  complex X[N], Y[N], D;
#pragma omp target data map(to:X[0 : N]) map(tofrom:Y[0 : N])
  zaxpy(X, Y, D, N);
}

This code is compiled using the following Clang flags.

$ clang++ -fopenmp -fopenmp-targets=nvptx64 -O3 zaxpy.cpp -c

The output section in the object file can be seen using the readelf utility

$ llvm-readelf -WS zaxpy.o
[Nr] Name                                       Type
...
[34] omp_offloading_entries                     PROGBITS
[35] .llvm.offloading.nvptx64-nvidia-cuda.sm_70 PROGBITS

Compiling this file again will invoke the clang-linker-wrapper utility to extract and link the device code stored at the section named .llvm.offloading.nvptx64-nvidia-cuda.sm_70 and then use entries stored in the section named omp_offloading_entries to create the symbols necessary for libomptarget to register the device image and call the entry function.

$ clang++ -fopenmp -fopenmp-targets=nvptx64 zaxpy.o -o zaxpy
$ ./zaxpy

We can see the steps created by clang to generate the offloading code using the -ccc-print-phases option in Clang. This matches the description in :ref:`Offloading Overview`.

$ clang++ -fopenmp -fopenmp-targets=nvptx64 -ccc-print-phases zaxpy.cpp
# "x86_64-unknown-linux-gnu" - "clang", inputs: ["zaxpy.cpp"], output: "/tmp/zaxpy-host.bc"
# "nvptx64-nvidia-cuda" - "clang", inputs: ["zaxpy.cpp", "/tmp/zaxpy-e6a41b.bc"], output: "/tmp/zaxpy-07f434.s"
# "nvptx64-nvidia-cuda" - "NVPTX::Assembler", inputs: ["/tmp/zaxpy-07f434.s"], output: "/tmp/zaxpy-0af7b7.o"
# "x86_64-unknown-linux-gnu" - "clang", inputs: ["/tmp/zaxpy-e6a41b.bc", "/tmp/zaxpy-0af7b7.o"], output: "/tmp/zaxpy-416cad.o"
# "x86_64-unknown-linux-gnu" - "Offload::Linker", inputs: ["/tmp/zaxpy-416cad.o"], output: "a.out"