Skip to content
Snippets Groups Projects
Commit f2a1844a authored by Pavel Jirásek's avatar Pavel Jirásek
Browse files

Merge branch 'remark' of gitlab.it4i.cz:it4i-admins/docs.it4i into remark

parents 0a67e6a4 d56fe278
No related branches found
No related tags found
No related merge requests found
Showing
with 42 additions and 191 deletions
...@@ -251,3 +251,7 @@ r37u31n1008 ...@@ -251,3 +251,7 @@ r37u31n1008
qsub qsub
it4ifree it4ifree
it4i.portal.clients it4i.portal.clients
API
GNU
CUDA
NVIDIA
# Anselm Cluster Software
## [Modules](../../modules-anselm)
- List of available modules
## [COMSOL](comsol-multiphysics)
- A finite element analysis, solver and Simulation software
## [ParaView](paraview)
- An open-source, multi-platform data analysis and visualization application
## [Compilers](compilers)
- Available compilers, including GNU, INTEL and UPC compilers
## [NVIDIA CUDA](nvidia-cuda)
- A guide to NVIDIA CUDA programming and GPU usage
## [GPI-2](gpi2)
- A library that implements the GASPI specification
## [OpenFOAM](openfoam)
- A free, open source CFD software package
## [ISV Licenses](isv_licenses)
- A guide to managing Independent Software Vendor licenses
## [Intel Xeon Phi](intel-xeon-phi)
- A guide to Intel Xeon Phi usage
## [Virtualization](kvirtualization)
## [Java](java)
- Java on ANSELM
## [Operating System](operating-system)
- The operating system, deployed on ANSELM
## Intel Suite
- The Intel Parallel Studio XE
### [Introduction](intel-suite/introduction)
### [Intel MKL](intel-suite/intel-mkl)
### [Intel Compilers](intel-suite/intel-compilers)
### [Intel IPP](intel-suite/intel-integrated-performance-primitives)
### [Intel TBB](intel-suite/intel-tbb)
### [Intel Debugger](intel-suite/intel-debugger)
## MPI
- Message Passing Interface libraries
### [Introduction](mpi/mpi)
### [MPI4Py (MPI for Python)](mpi/mpi4py-mpi-for-python)
### [Running OpenMPI](mpi/Running_OpenMPI)
### [Running MPICH2](mpi/running-mpich2)
## Numerical Libraries
- Libraries for numerical computations
### [Intel numerical libraries](numerical-libraries/intel-numerical-libraries)
### [PETSc](numerical-libraries/petsc)
### [Trilinos](numerical-libraries/trilinos)
### [FFTW](numerical-libraries/fftw)
### [GSL](numerical-libraries/gsl)
### [MAGMA for Intel Xeon Phi](numerical-libraries/magma-for-intel-xeon-phi)
### [HDF5](numerical-libraries/hdf5)
## Omics Master
### [Diagnostic component (TEAM)](omics-master/diagnostic-component-team)
### [Prioritization component (BiERapp)](omics-master/priorization-component-bierapp)
### [Overview](omics-master/overview)
## Debuggers
- A collection of development tools
### [Valgrind](debuggers/valgrind)
### [PAPI](debuggers/papi)
### [Allinea Forge (DDT,MAP)](debuggers/allinea-ddt)
### [Total View](debuggers/total-view)
### [CUBE](debuggers/cube)
### [Intel VTune Amplifier](debuggers/intel-vtune-amplifier)
### [VNC](debuggers/debuggers)
### [Scalasca](debuggers/scalasca)
### [Score-P](debuggers/score-p)
### [Intel Performance Counter Monitor](debuggers/intel-performance-counter-monitor)
### [Allinea Performance Reports](debuggers/allinea-performance-reports)
### [Vampir](debuggers/vampir)
## Numerical Languages
- Interpreted languages for numerical computations
### [Introduction](numerical-languages/introduction)
### [R](numerical-languages/r)
### [MATLAB 2013-2014](numerical-languages/matlab_1314)
### [MATLAB](numerical-languages/matlab)
### [Octave](numerical-languages/octave)
## Chemistry
- Tools for computational chemistry
### [Molpro](chemistry/molpro)
### [NWChem](chemistry/nwchem)
## ANSYS
- An engineering simulation software
### [Introduction](ansys/ansys)
### [ANSYS CFX](ansys/ansys-cfx)
### [ANSYS LS-DYNA](ansys/ansys-ls-dyna)
### [ANSYS MAPDL](ansys/ansys-mechanical-apdl)
### [LS-DYNA](ansys/ls-dyna)
### [ANSYS Fluent](ansys/ansys-fluent)
...@@ -6,7 +6,7 @@ Intel Math Kernel Library (Intel MKL) is a library of math kernel subroutines, e ...@@ -6,7 +6,7 @@ Intel Math Kernel Library (Intel MKL) is a library of math kernel subroutines, e
- BLAS (level 1, 2, and 3) and LAPACK linear algebra routines, offering vector, vector-matrix, and matrix-matrix operations. - BLAS (level 1, 2, and 3) and LAPACK linear algebra routines, offering vector, vector-matrix, and matrix-matrix operations.
- The PARDISO direct sparse solver, an iterative sparse solver, and supporting sparse BLAS (level 1, 2, and 3) routines for solving sparse systems of equations. - The PARDISO direct sparse solver, an iterative sparse solver, and supporting sparse BLAS (level 1, 2, and 3) routines for solving sparse systems of equations.
- ScaLAPACK distributed processing linear algebra routines for Linux_ and Windows_ operating systems, as well as the Basic Linear Algebra Communications Subprograms (BLACS) and the Parallel Basic Linear Algebra Subprograms (PBLAS). - ScaLAPACK distributed processing linear algebra routines for Linux and Windows operating systems, as well as the Basic Linear Algebra Communications Subprograms (BLACS) and the Parallel Basic Linear Algebra Subprograms (PBLAS).
- Fast Fourier transform (FFT) functions in one, two, or three dimensions with support for mixed radices (not limited to sizes that are powers of 2), as well as distributed versions of these functions. - Fast Fourier transform (FFT) functions in one, two, or three dimensions with support for mixed radices (not limited to sizes that are powers of 2), as well as distributed versions of these functions.
- Vector Math Library (VML) routines for optimized mathematical operations on vectors. - Vector Math Library (VML) routines for optimized mathematical operations on vectors.
- Vector Statistical Library (VSL) routines, which offer high-performance vectorized random number generators (RNG) for several probability distributions, convolution and correlation routines, and summary statistics functions. - Vector Statistical Library (VSL) routines, which offer high-performance vectorized random number generators (RNG) for several probability distributions, convolution and correlation routines, and summary statistics functions.
......
...@@ -2,13 +2,11 @@ ...@@ -2,13 +2,11 @@
The Anselm cluster provides following elements of the Intel Parallel Studio XE The Anselm cluster provides following elements of the Intel Parallel Studio XE
|Intel Parallel Studio XE| * Intel Compilers
\|-------------------------------------------------\| * Intel Debugger
|Intel Compilers| * Intel MKL Library
|Intel Debugger| * Intel Integrated Performance Primitives Library
|Intel MKL Library| * Intel Threading Building Blocks Library
|Intel Integrated Performance Primitives Library|
|Intel Threading Building Blocks Library|
## Intel compilers ## Intel compilers
......
# Intel Xeon Phi # Intel Xeon Phi
\##A guide to Intel Xeon Phi usage ## A guide to Intel Xeon Phi usage
Intel Xeon Phi can be programmed in several modes. The default mode on Anselm is offload mode, but all modes described in this document are supported. Intel Xeon Phi can be programmed in several modes. The default mode on Anselm is offload mode, but all modes described in this document are supported.
...@@ -9,19 +9,19 @@ Intel Xeon Phi can be programmed in several modes. The default mode on Anselm is ...@@ -9,19 +9,19 @@ Intel Xeon Phi can be programmed in several modes. The default mode on Anselm is
To get access to a compute node with Intel Xeon Phi accelerator, use the PBS interactive session To get access to a compute node with Intel Xeon Phi accelerator, use the PBS interactive session
```bash ```bash
$ qsub -I -q qmic -A NONE-0-0 $ qsub -I -q qmic -A NONE-0-0
``` ```
To set up the environment module "Intel" has to be loaded To set up the environment module "Intel" has to be loaded
```bash ```bash
$ module load intel/13.5.192 $ module load intel/13.5.192
``` ```
Information about the hardware can be obtained by running the micinfo program on the host. Information about the hardware can be obtained by running the micinfo program on the host.
```bash ```bash
$ /usr/bin/micinfo $ /usr/bin/micinfo
``` ```
The output of the "micinfo" utility executed on one of the Anselm node is as follows. (note: to get PCIe related details the command has to be run with root privileges) The output of the "micinfo" utility executed on one of the Anselm node is as follows. (note: to get PCIe related details the command has to be run with root privileges)
...@@ -93,14 +93,14 @@ The output of the "micinfo" utility executed on one of the Anselm node is as fol ...@@ -93,14 +93,14 @@ The output of the "micinfo" utility executed on one of the Anselm node is as fol
To compile a code for Intel Xeon Phi a MPSS stack has to be installed on the machine where compilation is executed. Currently the MPSS stack is only installed on compute nodes equipped with accelerators. To compile a code for Intel Xeon Phi a MPSS stack has to be installed on the machine where compilation is executed. Currently the MPSS stack is only installed on compute nodes equipped with accelerators.
```bash ```bash
$ qsub -I -q qmic -A NONE-0-0 $ qsub -I -q qmic -A NONE-0-0
$ module load intel/13.5.192 $ module load intel/13.5.192
``` ```
For debugging purposes it is also recommended to set environment variable "OFFLOAD_REPORT". Value can be set from 0 to 3, where higher number means more debugging information. For debugging purposes it is also recommended to set environment variable "OFFLOAD_REPORT". Value can be set from 0 to 3, where higher number means more debugging information.
```bash ```bash
export OFFLOAD_REPORT=3 export OFFLOAD_REPORT=3
``` ```
A very basic example of code that employs offload programming technique is shown in the next listing. Please note that this code is sequential and utilizes only single core of the accelerator. A very basic example of code that employs offload programming technique is shown in the next listing. Please note that this code is sequential and utilizes only single core of the accelerator.
...@@ -607,7 +607,7 @@ Intel MPI for the Xeon Phi coprocessors offers different MPI programming models: ...@@ -607,7 +607,7 @@ Intel MPI for the Xeon Phi coprocessors offers different MPI programming models:
**Symmetric model** - the MPI ranks reside on both the host and the coprocessor. Most general MPI case. **Symmetric model** - the MPI ranks reside on both the host and the coprocessor. Most general MPI case.
\###Host-only model ### Host-only model
In this case all environment variables are set by modules, so to execute the compiled MPI program on a single node, use: In this case all environment variables are set by modules, so to execute the compiled MPI program on a single node, use:
...@@ -809,7 +809,7 @@ The same way MPI program can be executed on multiple hosts: ...@@ -809,7 +809,7 @@ The same way MPI program can be executed on multiple hosts:
: -host cn205 -n 6 ~/mpi-test : -host cn205 -n 6 ~/mpi-test
``` ```
\###Symmetric model ### Symmetric model
In a symmetric mode MPI programs are executed on both host computer(s) and MIC accelerator(s). Since MIC has a different In a symmetric mode MPI programs are executed on both host computer(s) and MIC accelerator(s). Since MIC has a different
architecture and requires different binary file produced by the Intel compiler two different files has to be compiled before MPI program is executed. architecture and requires different binary file produced by the Intel compiler two different files has to be compiled before MPI program is executed.
......
# ISV Licenses # ISV Licenses
\##A guide to managing Independent Software Vendor licenses ## A guide to managing Independent Software Vendor licenses
On Anselm cluster there are also installed commercial software applications, also known as ISV (Independent Software Vendor), which are subjects to licensing. The licenses are limited and their usage may be restricted only to some users or user groups. On Anselm cluster there are also installed commercial software applications, also known as ISV (Independent Software Vendor), which are subjects to licensing. The licenses are limited and their usage may be restricted only to some users or user groups.
...@@ -57,9 +57,7 @@ Example of the Commercial Matlab license state: ...@@ -57,9 +57,7 @@ Example of the Commercial Matlab license state:
Each feature of each license is accounted and checked by the scheduler of PBS Pro. If you ask for certain licenses, the scheduler won't start the job until the asked licenses are free (available). This prevents to crash batch jobs, just because of unavailability of the needed licenses. Each feature of each license is accounted and checked by the scheduler of PBS Pro. If you ask for certain licenses, the scheduler won't start the job until the asked licenses are free (available). This prevents to crash batch jobs, just because of unavailability of the needed licenses.
The general format of the name is: The general format of the name is `feature__APP__FEATURE`.
**feature**APP**FEATURE**
Names of applications (APP): Names of applications (APP):
......
# Java # Java
\##Java on ANSELM ## Java on ANSELM
Java is available on Anselm cluster. Activate java by loading the java module Java is available on Anselm cluster. Activate java by loading the java module
......
# Virtualization # Virtualization
\##Running virtual machines on compute nodes Running virtual machines on compute nodes
## Introduction ## Introduction
...@@ -26,12 +26,13 @@ Virtualization has also some drawbacks, it is not so easy to setup efficient sol ...@@ -26,12 +26,13 @@ Virtualization has also some drawbacks, it is not so easy to setup efficient sol
Solution described in chapter [HOWTO](virtualization/#howto) is suitable for single node tasks, does not introduce virtual machine clustering. Solution described in chapter [HOWTO](virtualization/#howto) is suitable for single node tasks, does not introduce virtual machine clustering.
!!! Note "Note" !!! Note
Please consider virtualization as last resort solution for your needs. Please consider virtualization as last resort solution for your needs.
!!! Warning
Please consult use of virtualization with IT4Innovation's support. Please consult use of virtualization with IT4Innovation's support.
For running Windows application (when source code and Linux native application are not available) consider use of Wine, Windows compatibility layer. Many Windows applications can be run using Wine with less effort and better performance than when using virtualization. For running Windows application (when source code and Linux native application are not available) consider use of Wine, Windows compatibility layer. Many Windows applications can be run using Wine with less effort and better performance than when using virtualization.
## Licensing ## Licensing
......
...@@ -6,7 +6,7 @@ The OpenMPI programs may be executed only via the PBS Workload manager, by enter ...@@ -6,7 +6,7 @@ The OpenMPI programs may be executed only via the PBS Workload manager, by enter
### Basic usage ### Basic usage
!!! Note "Note" !!! Note
Use the mpiexec to run the OpenMPI code. Use the mpiexec to run the OpenMPI code.
Example: Example:
...@@ -27,7 +27,7 @@ Example: ...@@ -27,7 +27,7 @@ Example:
Hello world! from rank 3 of 4 on host cn110 Hello world! from rank 3 of 4 on host cn110
``` ```
!!! Note "Note" !!! Note
Please be aware, that in this example, the directive **-pernode** is used to run only **one task per node**, which is normally an unwanted behaviour (unless you want to run hybrid code with just one MPI and 16 OpenMP tasks per node). In normal MPI programs **omit the -pernode directive** to run up to 16 MPI tasks per each node. Please be aware, that in this example, the directive **-pernode** is used to run only **one task per node**, which is normally an unwanted behaviour (unless you want to run hybrid code with just one MPI and 16 OpenMP tasks per node). In normal MPI programs **omit the -pernode directive** to run up to 16 MPI tasks per each node.
In this example, we allocate 4 nodes via the express queue interactively. We set up the openmpi environment and interactively run the helloworld_mpi.x program. Note that the executable helloworld_mpi.x must be available within the In this example, we allocate 4 nodes via the express queue interactively. We set up the openmpi environment and interactively run the helloworld_mpi.x program. Note that the executable helloworld_mpi.x must be available within the
...@@ -48,7 +48,7 @@ You need to preload the executable, if running on the local scratch /lscratch fi ...@@ -48,7 +48,7 @@ You need to preload the executable, if running on the local scratch /lscratch fi
In this example, we assume the executable helloworld_mpi.x is present on compute node cn17 on local scratch. We call the mpiexec whith the **--preload-binary** argument (valid for openmpi). The mpiexec will copy the executable from cn17 to the /lscratch/15210.srv11 directory on cn108, cn109 and cn110 and execute the program. In this example, we assume the executable helloworld_mpi.x is present on compute node cn17 on local scratch. We call the mpiexec whith the **--preload-binary** argument (valid for openmpi). The mpiexec will copy the executable from cn17 to the /lscratch/15210.srv11 directory on cn108, cn109 and cn110 and execute the program.
!!! Note "Note" !!! Note
MPI process mapping may be controlled by PBS parameters. MPI process mapping may be controlled by PBS parameters.
The mpiprocs and ompthreads parameters allow for selection of number of running MPI processes per node as well as number of OpenMP threads per MPI process. The mpiprocs and ompthreads parameters allow for selection of number of running MPI processes per node as well as number of OpenMP threads per MPI process.
...@@ -97,7 +97,7 @@ In this example, we demonstrate recommended way to run an MPI application, using ...@@ -97,7 +97,7 @@ In this example, we demonstrate recommended way to run an MPI application, using
### OpenMP thread affinity ### OpenMP thread affinity
!!! Note "Note" !!! Note
Important! Bind every OpenMP thread to a core! Important! Bind every OpenMP thread to a core!
In the previous two examples with one or two MPI processes per node, the operating system might still migrate OpenMP threads between cores. You might want to avoid this by setting these environment variable for GCC OpenMP: In the previous two examples with one or two MPI processes per node, the operating system might still migrate OpenMP threads between cores. You might want to avoid this by setting these environment variable for GCC OpenMP:
...@@ -108,16 +108,16 @@ In the previous two examples with one or two MPI processes per node, the operati ...@@ -108,16 +108,16 @@ In the previous two examples with one or two MPI processes per node, the operati
or this one for Intel OpenMP: or this one for Intel OpenMP:
````bash ```bash
$ export KMP_AFFINITY=granularity=fine,compact,1,0 $ export KMP_AFFINITY=granularity=fine,compact,1,0
`` ```
As of OpenMP 4.0 (supported by GCC 4.9 and later and Intel 14.0 and later) the following variables may be used for Intel or GCC: As of OpenMP 4.0 (supported by GCC 4.9 and later and Intel 14.0 and later) the following variables may be used for Intel or GCC:
```bash ```bash
$ export OMP_PROC_BIND=true $ export OMP_PROC_BIND=true
$ export OMP_PLACES=cores $ export OMP_PLACES=cores
```` ```
## OpenMPI Process Mapping and Binding ## OpenMPI Process Mapping and Binding
...@@ -152,7 +152,7 @@ In this example, we see that ranks have been mapped on nodes according to the or ...@@ -152,7 +152,7 @@ In this example, we see that ranks have been mapped on nodes according to the or
Exact control of MPI process placement and resource binding is provided by specifying a rankfile Exact control of MPI process placement and resource binding is provided by specifying a rankfile
!!! Note "Note" !!! Note
Appropriate binding may boost performance of your application. Appropriate binding may boost performance of your application.
Example rankfile Example rankfile
......
...@@ -51,7 +51,7 @@ For example ...@@ -51,7 +51,7 @@ For example
comm.Barrier() # wait for everybody to synchronize comm.Barrier() # wait for everybody to synchronize
``` ```
\###Collective Communication with NumPy arrays ### Collective Communication with NumPy arrays
```cpp ```cpp
from mpi4py import MPI from mpi4py import MPI
......
# Salomon Cluster Software
## [Modules](../../modules-salomon)
- List of Available Modules
## [Compilers](compilers)
- Available compilers, including GNU, INTEL and UPC compilers
## [Intel Xeon Phi](intel-xeon-phi)
- A guide to Intel Xeon Phi usage
## [Java](java)
- Java on the cluster
## [Operating System](operating-system)
- The operating system, deployed on Salomon cluster
## Intel Suite
- The Intel Parallel Studio XE
### [Introduction](intel-suite/intel-parallel-studio-introduction)
### [Intel MKL](intel-suite/intel-mkl)
### [Intel Compilers](intel-suite/intel-compilers)
### [Intel IPP](intel-suite/intel-integrated-performance-primitives)
### [Intel TBB](intel-suite/intel-tbb)
### [Intel Debugger](intel-suite/intel-debugger)
### [Intel Inspector](intel-suite/intel-inspector)
### [Intel Trace Analyzer and Collector](intel-suite/intel-trace-analyzer-and-collector)
### [Intel Advisor](intel-suite/intel-advisor)
## MPI
- Message Passing Interface libraries
### [Introduction](mpi/mpi)
### [MPI4Py (MPI for Python)](mpi/mpi4py-mpi-for-python)
### [Running Open MPI](mpi/Running_OpenMPI)
## Debuggers
- A collection of development tools
### [Introduction](debuggers/Introduction)
### [Valgrind](debuggers/valgrind)
### [Allinea Forge (DDT,MAP)](debuggers/allinea-ddt)
### [Total View](debuggers/total-view)
### [Intel VTune Amplifier XE](debuggers/intel-vtune-amplifier)
### [Aislinn](debuggers/aislinn)
### [Allinea Performance Reports](debuggers/allinea-performance-reports)
### [Vampir](debuggers/vampir)
## Numerical Languages
- Interpreted languages for numerical computations
### [Introduction](numerical-languages/introduction)
### [R](numerical-languages/r)
### [MATLAB](numerical-languages/matlab)
### [Octave](numerical-languages/octave)
## Chemistry
- Tools for computational chemistry
### [Molpro](chemistry/molpro)
### [Phono3py](chemistry/phono3py)
### [NWChem](chemistry/nwchem)
## COMSOL
- A finite element analysis, solver and Simulation software
### [COMSOL](comsol/comsol-multiphysics)
### [Licensing and Available Versions](comsol/licensing-and-available-versions)
## ANSYS
- An engineering simulation software
### [Introduction](ansys/ansys)
### [Workbench](ansys/workbench)
### [ANSYS CFX](ansys/ansys-cfx)
### [ANSYS LS-DYNA](ansys/ansys-ls-dyna)
### [ANSYS MAPDL](ansys/ansys-mechanical-apdl)
### [ANSYS Fluent](ansys/ansys-fluent)
### [Setting license preferences](ansys/licensing)
### [Licensing and Available Versions](ansys/setting-license-preferences)
...@@ -35,7 +35,7 @@ def main(location): ...@@ -35,7 +35,7 @@ def main(location):
disabled = 0 disabled = 0
if line.startswith('#') and disabled == 0: if line.startswith('#') and disabled == 0:
line = titlecase(line[:(line.find("]"))], callback=abbreviations)+line[(line.find("]")):] line = titlecase(line[:(line.find("]"))], callback=abbreviations)+line[(line.find("]")):]
if line.startswith('---') or line.startswith('==='): if (line.startswith('---') or line.startswith('===')) and disabled == 0:
prev_line = titlecase(prev_line[:(prev_line.find("]"))], callback=abbreviations)+prev_line[(prev_line.find("]")):] prev_line = titlecase(prev_line[:(prev_line.find("]"))], callback=abbreviations)+prev_line[(prev_line.find("]")):]
f.write(prev_line) f.write(prev_line)
prev_line = line prev_line = line
......
...@@ -41,7 +41,7 @@ def main(location): ...@@ -41,7 +41,7 @@ def main(location):
print("-"+line,end="") print("-"+line,end="")
print("+"+titlecase(line[:(line.find("]"))], callback=abbreviations)+line[(line.find("]")):],end="") print("+"+titlecase(line[:(line.find("]"))], callback=abbreviations)+line[(line.find("]")):],end="")
return_value = 1 return_value = 1
if line.startswith('---') or line.startswith('==='): if (line.startswith('---') or line.startswith('===')) and disabled == 0:
if prev_line != titlecase(prev_line[:(prev_line.find("]"))], callback=abbreviations)+prev_line[(prev_line.find("]")):]: if prev_line != titlecase(prev_line[:(prev_line.find("]"))], callback=abbreviations)+prev_line[(prev_line.find("]")):]:
print() print()
print("-"+prev_line,end="") print("-"+prev_line,end="")
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment