Astrophysics ORION


Name of the cluster:

ORION

Institution:

Max Planck Institute for Astrophysics

Login nodes:

  • orion01.bc.mpcdf.mpg.de

  • orion02.bc.mpcdf.mpg.de

Their SHA256 ssh host key fingerprint is:

SHA256:1puEKbiOBuwsd4ak/5WQmL1NhfBysAUN1HJpnboghqE (ED25519)

Hardware Configuration:

  • login nodes orion0[12] : 2 x Intel(R) Xeon(R) Platinum 8480+ CPU @ 2.00GHz; 112 cores per node; 2 threads per core; 1 TB RAM

  • 104 execution nodes orion[001-104] for parallel computing with a total amount of 11648 CPU cores; 2 x Intel(R) Xeon(R) Platinum 8480+ CPU @ 2.00GHz; 512 GB RAM; Turbo-mode is on

  • node interconnect is based on Infiniband Fabric (Speed: 200Gb/s)

Filesystems:

/u

shared home filesystem; GPFS-based; user quotas (currently 100 GB, 1M files) enforced; quota can be checked with ‘/usr/lpp/mmfs/bin/mmlsquota’.

/orion/ptmp

shared scratch filesystem (5.0 PB); GPFS-based; user quotas (currently 20TB) enforced. NO BACKUPS!

/virgotng

shared scratch filesystem (8.0 PB); GPFS-based; no quotas enforced. NO BACKUPS!

Compilers and Libraries:

The “module” subsystem is implemented on ORION. Please use ‘module available’ to see all available modules.

  • Intel compilers (-> ‘module load intel’): icx, icpx, ifort

  • GNU compilers (-> ‘module load gcc’): gcc, g++, gfortran

  • Intel MKL (‘module load mkl’): $MKL_HOME defined; libraries found in $MKL_HOME/lib/intel64

  • Intel MPI 2024.0 (‘module load impi’): mpicc, mpigcc, mpiicc, mpiifort, mpiexec, …´

Similar to the HPC systems, this module tree is hierarchical.
To find a module and information about the available versions or what dependencies need to be loaded first one can use the ‘find-module’ command.

Batch system based on Slurm:

  • sbatch, srun, squeue, sinfo, scancel, scontrol, s*

  • current max. turnaround time (wallclock): 24 hours

  • two partition: p.exclusive (default) and p.shared (6 nodes) for shared resources

  • max. nodes limit per user in p.exclusive partition: 30

  • sample batch scripts can be found on Raven home page (must be modified for ORION)

Useful tips:

Nodes in p.shared partition are in shared mode, default memory per job set to 32000 MB. To allocate necessary amount of memory use --mem parameter.

Support:

For support please create a trouble ticket at the MPCDF helpdesk