Changes between Version 6 and Version 7 of Public/User_Guide/DEEP-EST_DAM


Ignore:
Timestamp:
Oct 15, 2019, 1:16:32 PM (5 years ago)
Author:
Jochen Kreutz
Comment:

JK: added info about FPGA usage

Legend:

Unmodified
Added
Removed
Modified
  • Public/User_Guide/DEEP-EST_DAM

    v6 v7  
    11= System usage =
    2 
    32The DEEP-EST Data Analytics Module (DAM) can be used through the SLURM based batch system that is also used for (most of) the Software Development Vehicles (SDV). You can request DAM nodes (`dp-dam[01-16]`) with an interactive session like this:
    43
     
    1514dp-dam04
    1615}}}
    17 
    18 
    1916When using a batch script, you have to adapt the partition option within your script: `--partition=dp-dam`
    2017
    2118== Using Cuda ==
    22 
    2319To compile and run Cuda applications on the Nvidia V100 cards included in the DAM nodes, it is necessary to load the CUDA module:
    2420
     
    3531}}}
    3632
     33== Using FPGAs ==
     34
     35Each node is equipped with a Stratix 10 FPGA. For getting started using OpenCL with the FPGAs you can find some hints as well as the slides and exercises from the Intel FPGA workshop held at JSC
     36
     37{{{
     38/usr/local/fpga
     39}}}
     40
     41It is recommended to do the first steps in an interactive session on a DAM node. To set up and check the FPGA environment, do the following:
     42
     43{{{
     44source /usr/local/fpga/FPGA_init.sh
     45lspci | grep -i 'accel'
     46aocl list-devices
     47aoc -list-boards
     48# -- optional for doing the exercises:
     49# export CL_CONTEXT_EMULATOR_DEVICE_INTELFPGA=1
     50}}}
     51
     52
     53You can copy and untar the lab into your home directory to do the exercises step by step. The exercises use the emulator device instead of the actual FPGA device due to the long
     54compilation time for the FPGAs. For using the FPGA device you will have to compile your OpenCL kernels using `-board=pac_s10_dc` option:
     55
     56{{{
     57# compile for the emulator
     58aoc -march=emulator -fast-emulator kernel-file.cl
     59
     60# compile for the FPGA device
     61aoc -board=pac_s10_dc kernel-file.cl
     62}}}
     63
     64In addition, you will have to adapt the OpenCL host file to select the correct platform ("Intel(R) FPGA SDK for OpenCL(TM)" or "Intel(R) FPGA Emulation Platform for OpenCL(TM) (preview)").
     65**Attention:** Compiling kernels for the FPGA device (instead of the emulator) might take several hours.
     66
     67Although eclipse is available on the DAM nodes, compiling and running the example applications might not work out, so you have to fall back to the command line as described in the exercise manual and by using
     68the provided `simple_compile.sh` scripts.
    3769
    3870== Filesystems and local storage ==
     71The home filesystem on the DEEP-EST Cluster Module is provided via GPFS/NFS and hence the same as on (most of) the remaining compute nodes.  The local storage system of the DAM running BeeGFS is available at
    3972
    40 The home filesystem on the DEEP-EST Cluster Module is provided via GPFS/NFS and hence the same as on (most of) the remaining compute nodes.
    41 The local storage system of the DAM running BeeGFS is available at
    4273{{{
    4374/work
     
    4576The file servers are reachable through the 40 GbE interface of the DAM nodes.
    4677
    47 This is NOT the same storage being used on the DEEP-ER SDV system. Both, the DEEP-EST prototype system and the DEEP-ER SDV have their own local storage.   
     78This is NOT the same storage being used on the DEEP-ER SDV system. Both, the DEEP-EST prototype system and the DEEP-ER SDV have their own local storage.
    4879
    4980It's possible to access the local storage of the DEEP-ER SDV (`/sdv-work`), but you have to keep in mind that the file servers of that storage can just be accessed through 1 GbE ! Hence, it should not be used for performance relevant applications since it is much slower than the DEEP-EST local storages mounted to `/work`.
     
    5182There is node local storage available for the DEEP-EST DAM node (2 x 1.5 TB NVMe SSD), but configuration is to be done for those devices.
    5283
    53 
    5484== Multi-node Jobs ==
    55 
    5685Currently, multi-node MPI jobs are possible on the DAM only by modifying the environment in the following way:
    5786
     
    6291Hello World from processor dp-dam01, rank 0 out of 2
    6392}}}
    64 
    6593**Attention:** This is a temporary workaround.
    6694
    67 {{{#!comment
    68 **Attention:** Since the Extoll network is not in place yet multi-node MPI Jobs are currently disabled.
     95{{{#!comment 
     96**Attention:** Since the Extoll network is not in place yet multi-node MPI Jobs are currently disabled. 
    6997}}}
    7098
    7199{{{#!comment
    72100
    73 Loading the most recent ParaStation module will be enough to run multi-node MPI jobs over Extoll 
     101Loading the most recent ParaStation module will be enough to run multi-node MPI jobs over Extoll
    74102
    75103{{{
    76104module load ParaStationMPI
    77105}}}
    78 
    79106}}}