-
Roman Sliva authoredRoman Sliva authored
karolina-slurm.md 1.69 KiB
Slurm Job Submission and Execution
Introduction
Slurm workload manager is used to allocate and access Karolina cluster's resources. General information about Slurm use at IT4Innovations can be found at Slurm Job Submission and Execution.
Getting Partition Information
Display partitions/queues on system:
$ sinfo -s
PARTITION AVAIL TIMELIMIT NODES(A/I/O/T) NODELIST
qcpu* up 2-00:00:00 0/720/0/720 cn[001-720]
qcpu_biz up 2-00:00:00 0/720/0/720 cn[001-720]
qcpu_exp up 1:00:00 0/720/0/720 cn[001-720]
qcpu_free up 18:00:00 0/720/0/720 cn[001-720]
qcpu_long up 6-00:00:00 0/720/0/720 cn[001-720]
qcpu_preempt up 12:00:00 0/720/0/720 cn[001-720]
qgpu up 2-00:00:00 0/71/1/72 acn[01-72]
qgpu_biz up 2-00:00:00 0/71/1/72 acn[01-72]
qgpu_exp up 1:00:00 0/71/1/72 acn[01-72]
qgpu_free up 18:00:00 0/71/1/72 acn[01-72]
qgpu_preempt up 12:00:00 0/71/1/72 acn[01-72]
qfat up 2-00:00:00 0/1/0/1 sdf1
qviz up 8:00:00 0/2/0/2 viz[1-2]
Graphical representation of cluster usage, partitions, nodes, and jobs could be found at [https://extranet.it4i.cz/rsweb/karolina][3]
On Karolina cluster
- all cpu queues/partitions provide full node allocation, whole nodes are allocated to job
- other queues/partitions (gpu, fat, viz) provide partial node allocation
Accessing CPU Queues
$ salloc -A PROJECT-ID -p qcpu