Changes between Version 16 and Version 17 of Public/User_Guide/Offloading_hybrid_apps


Ignore:
Timestamp:
Sep 17, 2019, 5:58:00 PM (5 years ago)
Author:
Kevin Sala
Comment:

Legend:

Unmodified
Added
Removed
Modified
  • Public/User_Guide/Offloading_hybrid_apps

    v16 v17  
    3838
    3939On the other hand, !OmpSs-2 proposes another approach targeting NVIDIA Unified Memory
    40 devices. CUDA kernels can be annotated as tasks declaring the corresponding data dependencies
    41 on the data buffers, so that, when all the dependencies of a task are satisfied, the CUDA
    42 kernel associated is offloaded to one of the available GPUs. In order to use that functionality,
    43 the user only has to allocate the buffers that CUDA kernels will access as a Unified Memory
    44 buffers (using the `cudaMallocManaged()` function).
     40devices. CUDA kernels can be annotated as regular tasks, and they can declare the
     41corresponding data dependencies on the data buffers. When all the dependencies of a CUDA task
     42are satisfied, the CUDA kernel associated to the task is offloaded to one of the available
     43GPUs. To use that functionality, the user only has to allocate the buffers that CUDA kernels
     44will access as Unified Memory buffers (i.e., using the `cudaMallocManaged()` function).
    4545
    4646== N-Body Benchmark ==