You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 15 Next »

Here we collect first user feedbacks with respect to albedos performance.

With/out Hyperthreading (SMT)

ModelUserPro SMT (smile)Contra SMT (sad)
idleadmin-

(∑ Esocket[0-7] according to lm_sensors:
Nodes need 30% more power (3500 kJ) and get warmer compared to without SMT (~2500 kJ)

stress-ng streamadmin--

~13% slower

FESOMNECUsing 128 Threads per node: 3% faster (probably because the (buggy) GXFS daemon can use a virtual core)Using 256 Thread per node: 10% slower
Python AIvhelmno impact/differenceno impact/difference
matlab
#SBATCH --cpus-per-task=16
vhelm
Runtime: 1440s instead of 1366s → ~5% slower


GPU nodes (A40 vs. A100)

ModelUserA40 vs. A100
 tensorflow-gpu AI applicationvhelmno difference

python3, matrix operations with with numpy (fat) vs cupy (gpu)

sviquera







Disk Access



albedoollie
Applicationusernode internal
/tmp (NVMe)
100 Gb Infiniband
/albedo (GPFS)
10 Gb Ethernet
/isibhv (NVMe)

node internal
/tmp (SSD)

100 Gb Omnipath
/work (BeeGFS)

10 Gb Ethernet
/isibhv (NVMe)

idl: reading 244 data filesvhelm~9 sec10~13 sec8~11 sec
spikes up to 181 sec
27~29 sec27~37 sec29~60 sec
spikes up to 98 sec
ls -f
ls # default with stat/color
directory with
30000 entires
0.08 sec
0.19 sec
0.04 sec
0.3 sec
0.03 sec
0.2 sec
0.1 sec
0.4 sec
0.2 sec
1.6 sec
0.08 sec
0.3~0.7 sec








  • ...

Runtime compared with ollie



albedo
GPFS
albedo
local NVMe
ollie
BeeGFS
idl vhelm

Cumulative time for loop and if conditions:  
Cumulative time file open: 
Cumulative time data read and file close: 
Total cumulative time: 
Total amount of imported data (Mega Bytes):   


0.32 s ( 3.11 %)
0.03 s ( 0.26 %)
9.95 s (96.62 %)
10.30 s
34269 (  3442 MB/s)


0.13 s ( 1.40 %)
0.01 s ( 0.06 %)
8.94 s (98.54 %)
9.07 s
34269 (  3832 MB/s)


3.48 s (12.73 %)
0.05 s ( 0.19 %)
23.77 s (87.07 %)
27.30 s
34269 (  1441 MB/s)


  • No labels