Changes between Version 39 and Version 40 of Public/User_Guide/PaS
- Timestamp:
- Aug 27, 2021, 11:47:07 AM (4 years ago)
Legend:
- Unmodified
- Added
- Removed
- Modified
-
Public/User_Guide/PaS
v39 v40 3 3 This page is intended to give a short overview on known issues and to provide potential solutions and workarounds to the issues seen. 4 4 5 ''Last update: 2021-0 4-09''5 ''Last update: 2021-08-27'' 6 6 7 [[span(style=color: #FF0000, System maintenance from Tuesday, 2021-04-13 to Thursday, 2021-04-15, no user access !)]]7 [[span(style=color: #FF0000, 2021-08-27: Filesystem issues, no user access !)]] 8 8 {{{#!comment highlighted red text 9 9 [[span(style=color: #FF0000, System maintenance from Monday, 2020-09-07 to Friday, 2020-09-11, no user access !)]] … … 17 17 === CM nodes === 18 18 19 * dp-cn05: memory issue - node at Megware for repair (#2682) 20 19 21 * dp-cn25: FW issues (#2495) 22 23 * dp-cn42: memory issue (#2675) 24 20 25 21 26 {{{#!comment 22 27 === DAM nodes === 23 28 … … 25 30 * dp-dam02: node currently reserved for special use case (#2554) 26 31 * dp-dam03: node currently reserved for special use case (#2242) 27 32 }}} 28 33 29 34 === ESB nodes === … … 33 38 }}} 34 39 35 * **dp-esb[01-25]: currently not avialable due to Fabri3 installation** 40 * dp-esb[01-25]: currently being prepared as rocky linux testbed 41 36 42 * dp-esb75: node currently reserved for special use case (#2568) 37 43 … … 39 45 === SDV nodes === 40 46 41 * several nodes have been taken offline:42 - deeper-sdv[11-16]43 - deeper storage system: (deeper-fs[01-03], deeper-raids)44 * deeper-sdv[01-10]: currently not available: configuration change needed (low priority) 47 * deeper-sdv cluster nodes (Haswell) have been taken offline: deeper-sdv[11-16] 48 - not included in SLURM anymore 49 - will be used for testing 50 45 51 * knl01: NVMe issues (#2011) 46 52 … … 50 56 === SLURM jobs === 51 57 52 - due to introduction of accounting with the start of the early access programthere is some re-configuration58 - due to introduction of accounting there is some re-configuration 53 59 of user accounts needed within SLURM to assign the correct QOS levels and priorities for the jobs 54 - this might lead to (temporary) failing job starts for certain users 55 - if you cannot start jobs via SLURM, please write an email to the support list: `sup(at)deep-est.eu` 56 57 58 === GPU direct usage with IB on ESB === 59 60 - only available via Developer stage, for testing load: 61 62 {{{ 63 ml --force purge 64 ml use $OTHERSTAGES 65 ml load Stages/Devel-2020 66 ml load Intel 67 ml load ParaStationMPI 68 }}} 69 70 - use `PSP_CUDA=1` and `PSP_UCP=1` 60 * this might lead to (temporary) failing job starts for certain users 61 * if you cannot start jobs via SLURM, please write an email to the support list: `sup(at)deep-est.eu` 71 62 72 63