Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

Table of Contents
outlinetrue
indent10px
absoluteUrltrue
exclude1

Login

Copy data from ollie

  1. login to ollie
  2. on ollie:

    Code Block
    languagebash
    rsync -Pauv --no-g /work/ollie/$USER/your-data  $USER@albedo0.dmawi.de:/albedo/work/projects/$YOURPROJECT/


Note: The other way round (rsync from albedo instead of from ollie) does not work, because of a specific route set on ollie.

Software

  • Albedo is running the operating system Rocky Linux release 8.6 (Green Obsidian).
  • Slurm 22.05 is used as the job scheduling system. Important details on its configuration on Albedo are given here: Slurm.
  • Details on the user software can be found here: Software.

Environment modules

On albedo we use environment modules to load/unload specific versions of software. Loading a module modifies environment variables so that the shell e.g. knows where to look for binaries.

You get an overview of all software installed by typing

Code Block
languagebash
module avail

To load and unload a module use

Code Block
languagebash
# load
module load <module>

# unload
module unload <loaded module>


Sometimes it might be useful to unload all loaded modules at once. This is done with

Code Block
languagebash
module purge


Usage of node's internal storage

All compute (including fat and gpu) nodes have a local NVMe disk mounted as /tmp. The GPU nodes have an additional storage /scratch. See System overview for the exact sizes.  We strongly encourage you to use these node-internal storage, which is faster than the global /albedo storage, if your job does lots of reading/writing. In particular, it might be beneficial to write your job output to the local disk and copy it to /albedo after your job is finished.

Code Block
languagebash
# Copy input data to the node, where your main MPI (rank 0) task runs
rsync -ur $INPUT_DATA /tmp/

# If you need the input data on every node, you have to add `srun` in front of the copy command
srun rsync -ur $INPUT_DATA /tmp/

# do the main calculation
srun $MY_GREAT_PROGRAM

# Copy your results from node where main MPI (rank 0) task runs to global storage
rsync -r /tmp/output/* /albedo/scratch/$MYPROJECT/output/


GPU monitoring

When using the GPUs you can monitor their usage with


Code Block
languagebash
ssh gpu-00[12]  # login
module load gpustat
gpustat -i1 --show-user --show-cmd -a



Code Block
languagebash
ssh gpu-00[12]  # login 
watch -d -n 0.51 nvidia-smi   # -d shows differences