Skip to content
Snippets Groups Projects
Commit 4525ac1f authored by Jan Siwiec's avatar Jan Siwiec
Browse files

Update slurmtopbs.md

parent 8c1993b3
No related branches found
No related tags found
1 merge request!343Slurmtopbs
Pipeline #23326 passed
# Migrating From Slurm # Migrating From SLURM
SLURM-optimized parallel jobs will not under PBS out of the box. SLURM-optimized parallel jobs will not under PBS out of the box.
Conversion to PBS standards is necessary. Here we provide hints on how to proceed. Conversion to PBS standards is necessary. Here we provide hints on how to proceed.
...@@ -12,7 +12,7 @@ The `PBS_O_WORKDIR` returns the directory, where the `qsub` command was submitte ...@@ -12,7 +12,7 @@ The `PBS_O_WORKDIR` returns the directory, where the `qsub` command was submitte
The `PBS_JOBID` returns the numercal identifyer of the job. The `PBS_JOBID` returns the numercal identifyer of the job.
The `qsub` always starts execution in the `$HOME` directory. The `qsub` always starts execution in the `$HOME` directory.
## Migrating PyTorch from SLURM ## Migrating PyTorch From SLURM
The Intel MPI provides some useful variables that may be used in the scripts executed via the MPI. The Intel MPI provides some useful variables that may be used in the scripts executed via the MPI.
these include `PMI_RANK`,`PMI_SIZE` and `MPI_LOCALRANKID`. these include `PMI_RANK`,`PMI_SIZE` and `MPI_LOCALRANKID`.
...@@ -21,6 +21,7 @@ these include `PMI_RANK`,`PMI_SIZE` and `MPI_LOCALRANKID`. ...@@ -21,6 +21,7 @@ these include `PMI_RANK`,`PMI_SIZE` and `MPI_LOCALRANKID`.
- The `PMI_SIZE` returns the process rank within the MPI_COMM_WORLD communicator - the number of processes - The `PMI_SIZE` returns the process rank within the MPI_COMM_WORLD communicator - the number of processes
For example: For example:
``` ```
$ mpirun -n 4 /bin/bash -c 'echo $PMI_SIZE' $ mpirun -n 4 /bin/bash -c 'echo $PMI_SIZE'
4 4
...@@ -28,6 +29,7 @@ $ mpirun -n 4 /bin/bash -c 'echo $PMI_SIZE' ...@@ -28,6 +29,7 @@ $ mpirun -n 4 /bin/bash -c 'echo $PMI_SIZE'
4 4
4 4
``` ```
In typical multi-gpu multi-node setting using PyTorch one needs to know: In typical multi-gpu multi-node setting using PyTorch one needs to know:
- World-size - i.e. the total number of GPUs in the system - World-size - i.e. the total number of GPUs in the system
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Please register or to comment