Attention
The resource partitions will likely change over time and this page will be updated accordingly.
Resource partitions
Viking’s resources are divided up into various partitions as layed out below, these may change over time as it becomes clear how best to share Viking’s resources.
Tip
To select a partition you use the --partition=
option in the jobscript, for example --partition=test
. The nodes
partition the default and will be used if no other partition is specified.
Partition |
Max Job |
Max Jobs |
Max Cores |
Max Running |
Default |
Default |
Number |
---|---|---|---|---|---|---|---|
nodes |
48 hrs |
No Limit |
960 |
5T |
1 |
5.2G |
114 |
week |
7 days |
No Limit |
576 |
3T |
1 |
5.2G |
12 |
month |
30 days |
No Limit |
192 |
1T |
1 |
5.2G |
2 |
test |
30 mins |
2 |
48 |
256G |
1 |
5.2G |
2 |
preempt |
30 days |
No Limit |
1 |
5.2G |
114 |
||
gpu |
3 days |
No Limit |
1 |
5.2G |
14 |
||
gpu_short |
30 mins |
3 |
1 |
5.2G |
1 |
||
gpu_week |
7 days |
3 |
1 |
5.2G |
14 |
||
gpu_interactive |
2 hrs |
1 |
1 |
5.2G |
1 |
||
gpuplus |
3 days |
No Limit |
1 |
5.2G |
6 |
||
himem |
2 days |
No Limit |
1 |
20G |
2 |
||
himem_week |
7 days |
No Limit |
1 |
20G |
1 |
||
interactive |
8 hours |
1 |
8 |
25G |
1 |
10G |
1 |
Additional partition information
- nodes
Each node has 96 cores.
Maximum limit of 960 cores per user.
- week
Partition for jobs that need to run for longer than the 48h max time allowed by the nodes partition, but for less than a week. There are 12 nodes in this partition and each user is limited to using a maximum of six node’s CPU and memory (576 cores and 3T of RAM).
- month
Partition for jobs that need to run for longer than the 7 day max time allowed by the week partition, but for less than 30 days. The max running cores and max running memory limits are practical limits, due to the resources available on the two nodes.
- test
For testing jobs to ensure they run as expected.
- preempt
Warning: Jobs submitted to this partition are subject to preemption. There is no guarantee when or if a job will run, or whether it will complete. This partition has no limits on the resources that can be requested, but a lower priority than the nodes partition. Jobs will run if there are free resources in the nodes partition, but will be cancelled if these resources are later required for another higher priority job.
- gpu
Partition for running jobs that require GPUs, see documentation for details about how to request GPUs.
Each of the 14 nodes house three nVidia A40 GPUs
Your job script must request at least one GPU (eg
#SBATCH --gres=gpu:1
)You are limited to no more than six GPUs at a time across all of your jobs running in the
gpu
partitionEach GPU is limited to a maximum of 32 CPU cores and 167G of memory per GPU (1/3 of the the node’s resources)
- gpu_short
Partition for running short jobs on a GPU
One dedicated node with three nVidia A40 GPUs
Your job script must request only one GPU (eg
#SBATCH --gres=gpu:1
) per jobPractical limit of three jobs at any one time, as the dedicated node only has three GPUs
Maximum memory per job is 167G
Maximum cores per job is 32
- gpu_week
Partition for running GPU jobs on any of the nVidia A40 nodes for up to a week
Maximum time is seven days (eg
#SBATCH --time=7-00:00:00
)Your job script should request only one GPU (eg
#SBATCH --gres=gpu:1
)The
gpu_week
partition is limited to running a maximum of three GPUs at any time, across all users
- gpu_interactive
Partition for running interactive jobs with a GPU
One dedicated node with three nVidia A40 GPUs
Your job script must request only one GPU (eg
#SBATCH --gres=gpu:1
)Only one job per user on this partition
Maximum memory per job is 167G
Maximum cores per job is 32
- gpuplus
Partition for running jobs that require more GPU power, see documentation for details about how to request GPUs request GPUs.
Each of the six nodes house two nVidia H100 GPUs
Your job script must request at least one GPU (eg
#SBATCH --gres=gpu:1
)You are limited to no more than two GPUs at a time across all of your jobs running in the
gpuplus
partition
- himem
For running jobs that require memory greater than that available in other partitions. Each of the two nodes (himem01 and himem02) have 96 cores. The max running cores and max running memory limits are practical limits, due to the resources available on the nodes.
- himem_week
Partition for running jobs that require memory greater than that available in other partitions. The max running cores and max running memory limits are practical limits, due to the resources available on the nodes.
- interactive
Partition for jobs that would like to use Viking interactively. This could be used for graphical work with the eight hour queue limit set to mimic a working day. You can run interactive jobs in the
nodes
partition (or any other), but the queue time will be shorter in theinteractive
partition.