Compute Partitions
There are several different Slurm partitions for different purposes and having different characteristics. The partitions are collected into the following groups based on what they provide:
Which partition to choose?
If you do not request a partition, your job will be placed in the default partition, which is the NHR standard96
partition for all users. This means non-NHR users will have their jobs pending indefinitely with the reason PartitionConfig
. Which partitions you have access to is listed on Types of User Accounts. For example new SCC users should specify --partition scc-cpu
instead for their non-GPU batch jobs.
Many partitions have variants with a suffix like :SUFFIX
appended to the (e.g. medium40:shared
).
The NAME:test
partitions are, as the name suggests, intended for shorter and smaller test runs.
These have a higher priority and a few dedicated nodes, but are limited in time and number of nodes.
For NHR users there are NAME:shared
partitions, where it is possible to use less than a full node, which is suitable for pre- and postprocessing. All non-NHR partitions are usually already shared and do not enforce a reservation of entire nodes.
A job running on a shared node is only accounted for its core/GPU/memory fraction it uses (the maximum of each resource is taken).
All non-shared nodes are exclusive to one job, which implies that full compute cost per node are paid, even if fewer resources were requested.
The available home/local-ssd/work/perm storages are discussed in Storage Systems.
An overview of all partitions and node statuses is provided by running
sinfo -r
To see detailed information about a node’s type, run
scontrol show node <nodename>