Skip to content
Snippets Groups Projects

Slurm Job Submission and Execution

Introduction

Slurm workload manager is used to allocate and access Karolina cluster's resources. General information about Slurm use at IT4Innovations can be found at Slurm Job Submission and Execution.

Getting Partition Information

Display partitions/queues on system:

$ sinfo -s
PARTITION    AVAIL  TIMELIMIT   NODES(A/I/O/T) NODELIST
qcpu*           up 2-00:00:00      0/720/0/720 cn[001-720]
qcpu_biz        up 2-00:00:00      0/720/0/720 cn[001-720]
qcpu_exp        up    1:00:00      0/720/0/720 cn[001-720]
qcpu_free       up   18:00:00      0/720/0/720 cn[001-720]
qcpu_long       up 6-00:00:00      0/720/0/720 cn[001-720]
qcpu_preempt    up   12:00:00      0/720/0/720 cn[001-720]
qgpu            up 2-00:00:00        0/71/1/72 acn[01-72]
qgpu_biz        up 2-00:00:00        0/71/1/72 acn[01-72]
qgpu_exp        up    1:00:00        0/71/1/72 acn[01-72]
qgpu_free       up   18:00:00        0/71/1/72 acn[01-72]
qgpu_preempt    up   12:00:00        0/71/1/72 acn[01-72]
qfat            up 2-00:00:00          0/1/0/1 sdf1
qviz            up    8:00:00          0/2/0/2 viz[1-2]

Graphical representation of cluster usage, partitions, nodes, and jobs could be found at [https://extranet.it4i.cz/rsweb/karolina][3]

On Karolina cluster

  • all cpu queues/partitions provide full node allocation, whole nodes are allocated to job
  • other queues/partitions (gpu, fat, viz) provide partial node allocation

Accessing CPU Queues

$ salloc -A PROJECT-ID -p qcpu

Using GPU Queues

Using Fat Queue

Using Viz Queue