diff --git a/docs.it4i/software/tools/ansys/workbench.md b/docs.it4i/software/tools/ansys/workbench.md index 59518f2ad6b86efced066356232b5899ce34da19..2f926d65f36116496056af6e92e9eaef4377a236 100644 --- a/docs.it4i/software/tools/ansys/workbench.md +++ b/docs.it4i/software/tools/ansys/workbench.md @@ -1,34 +1,39 @@ -!!!warning - This page has not been updated yet. The page does not reflect the transition from PBS to Slurm. - # Workbench ## Workbench Batch Mode -It is possible to run Workbench scripts in a batch mode. You need to configure solvers of individual components to run in parallel mode. Open your project in Workbench. Then, for example, in *Mechanical*, go to *Tools - Solve Process Settings...*. +It is possible to run Workbench scripts in a batch mode. +You need to configure solvers of individual components to run in parallel mode. +Open your project in Workbench. +Then, for example, in *Mechanical*, go to *Tools - Solve Process Settings...*.  -Enable the *Distribute Solution* checkbox and enter the number of cores (e.g. 72 to run on two Barbora nodes). If you want the job to run on more than 1 node, you must also provide a so called MPI appfile. In the *Additional Command Line Arguments* input field, enter: +Enable the *Distribute Solution* checkbox and enter the number of cores (e.g. 72 to run on two Barbora nodes). +If you want the job to run on more than 1 node, you must also provide a so called MPI appfile. +In the *Additional Command Line Arguments* input field, enter: ```console -mpifile /path/to/my/job/mpifile.txt ``` -Where /path/to/my/job is the directory where your project is saved. We will create the file mpifile.txt programmatically later in the batch script. For more information, refer to \*ANSYS Mechanical APDL Parallel Processing\* \*Guide\*. +Where `/path/to/my/job` is the directory where your project is saved. +We will create the file `mpifile.txt` programmatically later in the batch script. +For more information, refer to \*ANSYS Mechanical APDL Parallel Processing\* \*Guide\*. -Now, save the project and close Workbench. We will use this script to launch the job: +Now, save the project and close Workbench. +We will use this script to launch the job: ```bash #!/bin/bash - #PBS -l select=2:ncpus=128 - #PBS -q qprod - #PBS -N test9_mpi_2 - #PBS -A OPEN-0-0 - + #SBATCH --nodes=2 + #SBATCH --ntasks-per-node=128 + #SBATCH --job-name=test9_mpi_2 + #SBATCH --partition=qcpu + #SBATCH --account=ACCOUNT_ID # change the working directory - DIR=/scratch/project/PROJECT_ID/$PBS_JOBID + DIR=/scratch/project/PROJECT_ID/$SLURM_JOB_ID mkdir -p "$DIR" cd "$DIR" || exit @@ -36,15 +41,15 @@ Now, save the project and close Workbench. We will use this script to launch the echo Time is `date` echo Directory is `pwd` echo This jobs runs on the following nodes: - echo `cat $PBS_NODEFILE` + echo `$SLURM_NODELIST` - ml ANSYS/21.1-intel-2018a + ml ANSYS/2023R2-intel-2022.12 #### Set number of processors per host listing procs_per_host=24 #### Create MPI appfile echo -n "" > mpifile.txt - for host in `cat $PBS_NODEFILE` + for host in `$SLURM_NODELIST` do echo "-h $host -np $procs_per_host $ANSYS160_DIR/bin/ansysdis161 -dis" > mpifile.txt done @@ -61,4 +66,6 @@ Now, save the project and close Workbench. We will use this script to launch the runwb2 -R jou6.wbjn -B -F test9.wbpj ``` -The solver settings are saved in the solvehandlers.xml file, which is not located in the project directory. Verify your solved settings when uploading a project from your local computer. +The solver settings are saved in the `solvehandlers.xml` file, +which is not located in the project directory. +Verify your solved settings when uploading a project from your local computer.