diff --git a/docs.it4i/software/intel/intel-xeon-phi-salomon.md b/docs.it4i/software/intel/intel-xeon-phi-salomon.md
index dde796968e9abea7eb5ca34770592f38c41a8389..3660a0ddb0d51dcfdd20968661d1ac53fecef48a 100644
--- a/docs.it4i/software/intel/intel-xeon-phi-salomon.md
+++ b/docs.it4i/software/intel/intel-xeon-phi-salomon.md
@@ -12,7 +12,7 @@ To get access to a compute node with the Intel Xeon Phi accelerator, use the PBS
 $ qsub -I -q qprod -l select=1:ncpus=24:accelerator=True:naccelerators=2:accelerator_model=phi7120 -A NONE-0-0
 ```
 
-To set up the environment load the "intel" module. Without specifying the version, the default version is loaded (at time of writing this, it is 2015b):
+To set up the environment load the `intel` module. Without specifying the version, the default version is loaded (at time of writing this, it is 2015b):
 
 ```console
 $ ml intel
@@ -146,7 +146,7 @@ $ qsub -I -q qprod -l select=1:ncpus=24:accelerator=True:naccelerators=2:acceler
 $ ml intel
 ```
 
-For debugging purposes, it is also recommended to set the "OFFLOAD_REPORT" environment variable. The value can be set from 0 to 3, where a higher number means more debugging information.
+For debugging purposes, it is also recommended to set the `OFFLOAD_REPORT` environment variable. The value can be set from 0 to 3, where a higher number means more debugging information.
 
 ```console
 export OFFLOAD_REPORT=3
@@ -174,7 +174,7 @@ int main(int argc, char* argv[])
 }
 ```
 
-To compile the code using Intel compiler, run:
+To compile the code using the Intel compiler, run:
 
 ```console
 $ icc source-offload.cpp -o bin-offload
@@ -268,7 +268,7 @@ int main()
 }
 ```
 
-During the compilation, the Intel compiler shows which loops have been vectorized in both the host and the accelerator. This can be enabled with the "-vec-report2" compiler option. To compile and execute the code, run:
+During the compilation, the Intel compiler shows which loops have been vectorized in both the host and the accelerator. This can be enabled with the `-vec-report2` compiler option. To compile and execute the code, run:
 
 ```console
 $ icc vect-add.c -openmp_report2 -vec-report2 -o vect-add
@@ -279,11 +279,11 @@ Some interesting compiler flags useful not only for code debugging are:
 
 !!! note
     Debugging
-    openmp_report[0|1|2] - controls the compiler based vectorization diagnostic level
-    vec-report[0|1|2] - controls the OpenMP parallelizer diagnostic level
+    `openmp_report[0|1|2]` - controls the compiler based vectorization diagnostic level
+    `vec-report[0|1|2]` - controls the OpenMP parallelizer diagnostic level
 
     Performance optimization
-    xhost - FOR HOST ONLY - to generate AVX (Advanced Vector Extensions) instructions.
+    `xhost` - FOR HOST ONLY - to generate AVX (Advanced Vector Extensions) instructions.
 
 ## Automatic Offload Using Intel MKL Library
 
@@ -308,7 +308,7 @@ To get more information about the automatic offload, refer to the "[Using Intel
 
 ### Automatic Offload Example
 
-At first, get an interactive PBS session on a node with the MIC accelerator and load the "intel" module that automatically loads the "mkl" module as well:
+At first, get an interactive PBS session on a node with the MIC accelerator and load the `intel` module that automatically loads the `mkl` module, as well:
 
 ```console
 $ qsub -I -q qprod -l select=1:ncpus=24:accelerator=True:naccelerators=2:accelerator_model=phi7120 -A NONE-0-0
@@ -410,12 +410,12 @@ Computing SGEMM with automatic workdivision
 Done
 ```
 
-!!! note ""
+!!! note
     Behavior of the automatic offload mode is controlled by functions called within the program or by environmental variables. The complete list of controls is listed [here][d].
 
 ### Automatic Offload Example #2
 
-In this example, we will demonstrate the automatic offload control via the MKL_MIC_ENABLE environment variable. The DGEMM function will be offloaded.
+In this example, we will demonstrate the automatic offload control via the `MKL_MIC_ENABLE` environment variable. The DGEMM function will be offloaded.
 
 At first, get an interactive PBS session on the node with the MIC accelerator.
 
@@ -460,9 +460,9 @@ $ ml intel
 ```
 
 !!! note
-    A particular version of the Intel module is specified. This information is used later to specify the correct library paths.
+    A particular version of the `intel` module is specified. This information is used later to specify the correct library paths.
 
-To produce a binary compatible with the Intel Xeon Phi architecture, the user has to specify the "-mmic" compiler flag. Two compilation examples are shown below. The first example shows how to compile the OpenMP parallel code "vect-add.c" for the host only:
+To produce a binary compatible with the Intel Xeon Phi architecture, the user has to specify the `-mmic` compiler flag. Two compilation examples are shown below. The first example shows how to compile the OpenMP parallel code `vect-add.c` for the host only:
 
 ```console
 $ icc -xhost -no-offload -fopenmp vect-add.c -o vect-add-host
@@ -484,7 +484,7 @@ $ icc -mmic -fopenmp vect-add.c -o vect-add-mic
 
 User access to Intel Xeon Phi is via SSH. Since user home directories are mounted using NFS on the accelerator, users do not have to copy binary files or libraries between the host and the accelerator.
 
-Get the PATH of MIC enabled libraries for currently used Intel Compiler (here icc/2015.3.187-GNU-5.1.0-2.25 was used):
+Get the PATH of MIC enabled libraries for currently used Intel Compiler (here `icc/2015.3.187-GNU-5.1.0-2.25` was used):
 
 ```console
 $ echo $MIC_LD_LIBRARY_PATH
@@ -503,14 +503,14 @@ If the code is sequential, it can be executed directly:
 mic0 $ ~/path_to_binary/vect-add-seq-mic
 ```
 
-If the code is parallelized using OpenMP, a set of additional libraries is required for execution. To locate these libraries, a new path has to be added to the LD_LIBRARY_PATH environment variable prior to the execution:
+If the code is parallelized using OpenMP, a set of additional libraries is required for execution. To locate these libraries, a new path has to be added to the `LD_LIBRARY_PATH` environment variable prior to the execution:
 
 ```console
 mic0 $ export LD_LIBRARY_PATH=/apps/all/icc/2015.3.187-GNU-5.1.0-2.25/composer_xe_2015.3.187/compiler/lib/mic:$LD_LIBRARY_PATH
 ```
 
 !!! note
-    The path exported in the previous example contains the path to a specific compiler (here the version is 2015.3.187-GNU-5.1.0-2.25). This version number has to match with the version number of the Intel compiler module that was used to compile the code on the host computer.
+    The path exported in the previous example contains the path to a specific compiler (here the version is `2015.3.187-GNU-5.1.0-2.25`). This version number has to match with the version number of the Intel compiler module that was used to compile the code on the host computer.
 
 For your information, the list of libraries and their location required for execution of an OpenMP parallel code on Intel Xeon Phi is:
 
@@ -535,7 +535,7 @@ On Salomon, OpenCL is installed only on compute nodes with the MIC accelerator,
 ml opencl-sdk opencl-rt
 ```
 
-Always load "opencl-sdk" (providing devel files like headers) and "opencl-rt" (providing a dynamic library libOpenCL.so) modules to compile and link OpenCL code. Load "opencl-rt" for running your compiled code.
+Always load `opencl-sdk` (providing devel files like headers) and `opencl-rt` (providing a dynamic library libOpenCL.so) modules to compile and link OpenCL code. Load `opencl-rt` for running your compiled code.
 
 There are two basic examples of OpenCL code in the following directory:
 
@@ -628,8 +628,8 @@ This ensures that MPI inside the node will use the SHMEM communication; between
 
 !!! note
     Other FABRICS like tcp or ofa may be used (even combined with shm) but there is a severe loss of performance (by order of magnitude).
-    Usage of a single DAPL PROVIDER (e.g. I_MPI_DAPL_PROVIDER=ofa-v2-mlx4_0-1u) will cause a failure of Host<->Phi and/or Phi<->Phi communication.
-    Usage of the I_MPI_DAPL_PROVIDER_LIST on a non-accelerated node will cause a failure of any MPI communication, since those nodes do not have aSCIF device and there's no CCL-Direct proxy runnig.
+    Usage of a single DAPL PROVIDER (e.g. `I_MPI_DAPL_PROVIDER=ofa-v2-mlx4_0-1u`) will cause a failure of Host<->Phi and/or Phi<->Phi communication.
+    Usage of the `I_MPI_DAPL_PROVIDER_LIST` on a non-accelerated node will cause a failure of any MPI communication, since those nodes do not have aSCIF device and there's no CCL-Direct proxy runnig.
 
 Again, an MPI code for Intel Xeon Phi has to be compiled on a compute node with the accelerator and MPSS software stack installed. To get to a compute node with the accelerator, use:
 
@@ -725,7 +725,7 @@ coprocessor; or 2) launch the task using "**mpiexec.hydra**" from a host.
 
 Similarly to execution of OpenMP programs in native mode, since the environmental module is not supported on MIC, the user has to setup paths to Intel MPI libraries and binaries manually. One-time setup can be done by creating a "**.profile**" file in the user's home directory. This file sets up the environment on the MIC automatically once the user accesses the accelerator via SSH.
 
-At first, get the LD_LIBRARY_PATH for currently used Intel Compiler and Intel MPI:
+At first, get the `LD_LIBRARY_PATH` for currently used Intel Compiler and Intel MPI:
 
 ```console
 $ echo $MIC_LD_LIBRARY_PATH
@@ -783,7 +783,7 @@ Hello world from process 0 of 4 on host r38u31n1000-mic0
 
 If the MPI program is launched from the host instead of the coprocessor, the environmental variables are not set using the ".profile" file. Therefore, the user has to specify library paths from the command line when calling "mpiexec".
 
-First step is to tell mpiexec that the MPI should be executed on the local accelerator by setting up the environmental variable "I_MPI_MIC"
+First step is to tell mpiexec that the MPI should be executed on the local accelerator by setting up the environmental variable `I_MPI_MIC`:
 
 ```console
 $ export I_MPI_MIC=1
@@ -795,7 +795,7 @@ Now the MPI program can be executed as:
 $ mpirun -genv LD_LIBRARY_PATH $MIC_LD_LIBRARY_PATH -host mic0 -n 4 ~/mpi-test-mic
 ```
 
-or using mpirun
+or using `mpirun`:
 
 ```console
 $ mpirun -genv LD_LIBRARY_PATH $MIC_LD_LIBRARY_PATH -host mic0 -n 4 ~/mpi-test-mic
@@ -803,7 +803,7 @@ $ mpirun -genv LD_LIBRARY_PATH $MIC_LD_LIBRARY_PATH -host mic0 -n 4 ~/mpi-test-m
 
 !!! note
     \* the full path to the binary has to be specified (here: "**>~/mpi-test-mic**")
-    \* the LD_LIBRARY_PATH has to match with the Intel MPI module used to compile the MPI code
+    \* the `LD_LIBRARY_PATH` has to match with the Intel MPI module used to compile the MPI code
 
 The output should be again similar to:
 
@@ -854,7 +854,7 @@ This output means that the PBS allocated nodes cn204 and cn205, which means that
     - to connect to the accelerator on the first node from the first node:  `$ ssh r25u25n710-mic0` or `$ ssh mic0`
     - to connect to the accelerator on the second node from the first node: `$ ssh r25u25n711-mic0`
 
-At this point, we expect that the correct modules are loaded and the binary is compiled. For parallel execution, the mpiexec.hydra is used. Again the first step is to tell mpiexec that the MPI can be executed on MIC accelerators by setting up the environmental variable "I_MPI_MIC"; do not forget to have correct FABRIC and PROVIDER defined.
+At this point, we expect that the correct modules are loaded and the binary is compiled. For parallel execution, the mpiexec.hydra is used. Again the first step is to tell mpiexec that the MPI can be executed on MIC accelerators by setting up the environmental variable `I_MPI_MIC`; do not forget to have correct FABRIC and PROVIDER defined.
 
 ```console
 $ export I_MPI_MIC=1
@@ -926,7 +926,7 @@ Hello world from process 2 of 4 on host r38u32n1001-mic0
 Hello world from process 3 of 4 on host r38u32n1001-mic0
 ```
 
-The execution procedure can be simplified by using the mpirun command with the machine file as a parameter. The machine file contains a list of all nodes and accelerators that should be used to execute MPI processes.
+The execution procedure can be simplified by using the `mpirun` command with the machine file as a parameter. The machine file contains a list of all nodes and accelerators that should be used to execute MPI processes.
 
 An example of a machine file that uses 2 >hosts (**r38u32n1001** and **r38u32n1002**) and 2 accelerators **(r38u32n1001-mic0** and **r38u32n1002-mic0**) to run 2 MPI processes on each of them:
 
@@ -938,7 +938,7 @@ r38u33n1002:2
 r38u33n1002-mic0:2
 ```
 
-In addition, if a naming convention is set in a way that the name of the binary for the host is **"bin_name"** and the name of the binary for the accelerator is **"bin_name-mic"** then by setting up the environment variable **I_MPI_MIC_POSTFIX** to **"-mic"**, the user does not have to specify the names of both binaries. In this case, mpirun needs just the name of the host binary file (i.e. "mpi-test") and uses the suffix to get the name of the binary for accelerator (i.e. "mpi-test-mic").
+In addition, if a naming convention is set in a way that the name of the binary for the host is **"bin_name"** and the name of the binary for the accelerator is **"bin_name-mic"** then by setting up the environment variable `I_MPI_MIC_POSTFIX` to `-mic`, the user does not have to specify the names of both binaries. In this case, `mpirun` needs just the name of the host binary file (i.e. "mpi-test") and uses the suffix to get the name of the binary for accelerator (i.e. "mpi-test-mic").
 
 ```console
 $ export I_MPI_MIC_POSTFIX=-mic
@@ -983,7 +983,7 @@ A set of node-files, which can be used instead of manually creating a new one ev
      - /lscratch/${PBS_JOBID}/nodefile-mic-sn MICs only node-file, using short names
      - /lscratch/${PBS_JOBID}/nodefile-mix-sn Hosts and MICs node-file, using short names
 
-Each host or accelerator is listed only once per file. The user has to specify how many jobs should be executed per node using the `-n` parameter of the mpirun command.
+Each host or accelerator is listed only once per file. The user has to specify how many jobs should be executed per node using the `-n` parameter of the `mpirun` command.
 
 ## Optimization