Important
This documentation covers the new GABA cluster, available since November 27th, 2025.
Brain Research
- Name of the clusters:
GABA
- Institution:
Max Planck Institute for Brain Research
Login nodes:
|
Hardware Configuration:
Login nodes gaba[11-15]:
CPUs Model : AuthenticAMD EPYC 9554 64-Core Processor
2 sockets per node; 64 cores per socket; hyper-threading is on
RAM : 768 GB
GPUs : 4 Nvidia L40s-48GB GPUs per node
29 execution nodes gabag[201-229] for parallel CPU and GPU computing (3712 CPU cores) :
CPUs Model : AuthenticAMD EPYC 9554 64-Core Processor
2 sockets per node; 64 cores per socket; hyper-threading is on
RAM : 768 GB
GPUs : 4 Nvidia L40s-48GB GPUs per node
Node interconnect is based on 25 Gb/s Ethernet
Filesystems:
- /u
shared home filesystem with user home directory in
/u/<username>; user quotas enforced (200GB data, 512K files)- /oldgaba
former $HOME filesystem (formerly known as /gaba ), read-only, only available on login nodes. NO BACKUPS anylonger!
- /gabaghi
filesystem for archive (458 TB); only available on login nodes
- /conndata
shared filesystem filesystem; no quotas enforced. NO BACKUPS!
- /connscratch
shared filesystem filesystem; no quotas enforced. NO BACKUPS!
- /tmpscratch
shared filesystem filesystem; no quotas enforced. NO BACKUPS!
- /wKlive
filesystem for live wKcubes project; only available on login nodes NO BACKUPS!
- /nexus/posix0
only available on gaba14 for data transfers to/from HPC systems
Compilers and Libraries:
The “module” subsystem is implemented on GABA. Please use ‘module available’ to see all available modules.
Batch system based on Slurm:
login nodes: gaba[11-15]; batch nodes: gabag[201-229]
sbatch, srun, squeue, sinfo, scancel, scontrol, s*
current default turnaround time (wallclock) is 24 hours and max. turnaround time is 100 hours
Useful tips (slurm part of cluster):
By default run time limit used for jobs that don’t specify a value is 24 hours. Use --time option for sbatch/srun to set a limit on the total run time of the job allocation but not longer than 100 hours
Default memory per node is 50G. To grant the job access to all of the memory on each node use --mem=0 option for sbatch/srun
The OpenMP codes require a variable OMP_NUM_THREADS to be set. This can be obtained from the Slurm environment variable $SLURM_CPUS_PER_TASK which is set when --cpus-per-task is specified in a sbatch script.
To check node features, general resources and scheduling weight of nodes use sinfo -O nodelist,features,gres,weight
Support:
For support please create a trouble ticket at the MPCDF helpdesk