diff --git a/docs.it4i/prace.md b/docs.it4i/prace.md
index 99022ed6ebc300e38884c820f04069802a61d1eb..92dc5fb3368d8ff21b66ea57a9d13a879a8327d4 100644
--- a/docs.it4i/prace.md
+++ b/docs.it4i/prace.md
@@ -2,27 +2,27 @@
 
 ## Introduction
 
-PRACE users coming to the TIER-1 systems offered through the DECI calls are in general treated as standard users and so most of the general documentation applies to them as well. This section shows the main differences for quicker orientation, but often uses references to the original documentation. PRACE users who don't undergo the full procedure (including signing the IT4I AuP on top of the PRACE AuP) will not have a password and thus access to some services intended for regular users. This can lower their comfort, but otherwise they should be able to use the TIER-1 system as intended. See the [Obtaining Login Credentials section][1], if the same level of access is required.
+PRACE users coming to the TIER-1 systems offered through the DECI calls are, in general, treated as standard users, so most of the general documentation applies to them as well. This section shows the main differences for quicker orientation, but often uses references to the original documentation. PRACE users who do not undergo the full procedure (including signing the IT4I AuP on top of the PRACE AuP) will not have a password and thus an access to some services intended for regular users. However, even with the limited access, they should be able to use the TIER-1 system as intended. If the same level of access is required, see the [Obtaining Login Credentials][1] section.
 
 All general [PRACE User Documentation][a] should be read before continuing reading the local documentation here.
 
 ## Help and Support
 
-If you have any troubles, need information, request support or want to install additional software, use [PRACE Helpdesk][b].
+If you need any information, request support, or want to install additional software, use the [PRACE Helpdesk][b].
 
-Information about the local services are provided in the [introduction of general user documentation Salomon][2] and [introduction of general user documentation Anselm][3]. Keep in mind, that standard PRACE accounts don't have a password to access the web interface of the local (IT4Innovations) request tracker and thus a new ticket should be created by sending an e-mail to support[at]it4i.cz.
+Information about the local services are provided in the [introduction of general user documentation Salomon][2] and [introduction of general user documentation Anselm][3]. Keep in mind, that standard PRACE accounts don't have a password to access the web interface of the local (IT4Innovations) request tracker and thus a new ticket should be created by sending an email to support[at]it4i.cz.
 
 ## Obtaining Login Credentials
 
-In general PRACE users already have a PRACE account setup through their HOMESITE (institution from their country) as a result of rewarded PRACE project proposal. This includes signed PRACE AuP, generated and registered certificates, etc.
+In general, PRACE users already have a PRACE account set up through their HOMESITE (institution from their country) as a result of a rewarded PRACE project proposal. This includes signed PRACE AuP, generated and registered certificates, etc.
 
-If there's a special need a PRACE user can get a standard (local) account at IT4Innovations. To get an account on a cluster, the user needs to obtain the login credentials. The procedure is the same as for general users of the cluster, so see the corresponding [section of the general documentation here][1].
+If there is a special need, a PRACE user can get a standard (local) account at IT4Innovations. To get an account on a cluster, the user needs to obtain the login credentials. The procedure is the same as for general users of the cluster, see the corresponding [section of the general documentation here][1].
 
 ## Accessing the Cluster
 
 ### Access With GSI-SSH
 
-For all PRACE users the method for interactive access (login) and data transfer based on grid services from Globus Toolkit (GSI SSH and GridFTP) is supported.
+For all PRACE users, the method for interactive access (login) and data transfer based on grid services from Globus Toolkit (GSI SSH and GridFTP) is supported.
 
 The user will need a valid certificate and to be present in the PRACE LDAP (contact your HOME SITE or the primary investigator of your project for LDAP account creation).
 
@@ -34,25 +34,25 @@ Most of the information needed by PRACE users accessing the TIER-1 systems can b
 * [Data transfer with GridFTP][f]
 * [Data transfer with gtransfer][g]
 
-Before you start to use any of the services don't forget to create a proxy certificate from your certificate:
+Before you start using any of the services, do not forget to create a proxy certificate from your certificate:
 
 ```console
 $ grid-proxy-init
 ```
 
-To check whether your proxy certificate is still valid (by default it's valid 12 hours), use:
+To check whether your proxy certificate is still valid (12 hours by default), use:
 
 ```console
 $ grid-proxy-info
 ```
 
-To access the cluster, several login nodes running GSI SSH service are available. The service is available from public Internet as well as from the internal PRACE network (accessible only from other PRACE partners).
+To access the cluster, several login nodes running the GSI SSH service are available. The service is available from public Internet as well as from the internal PRACE network (accessible only from other PRACE partners).
 
 #### Access From PRACE Network:
 
-It is recommended to use the single DNS name **name-cluster**-prace.it4i.cz which is distributed between the four login nodes. If needed, user can login directly to one of the login nodes. The addresses are:
+It is recommended to use the single DNS name **name-cluster**-prace.it4i.cz which is distributed between the four login nodes. If needed, the user can log in directly to one of the login nodes. The addresses are:
 
-For Salomon cluster:
+Salomon cluster:
 
 | Login address                | Port | Protocol | Login node                       |
 | ---------------------------- | ---- | -------- | -------------------------------- |
@@ -66,7 +66,7 @@ For Salomon cluster:
 $ gsissh -p 2222 salomon-prace.it4i.cz
 ```
 
-For Anselm cluster:
+Anselm cluster:
 
 | Login address               | Port | Protocol | Login node       |
 | --------------------------- | ---- | -------- | ---------------- |
@@ -90,9 +90,9 @@ $ gsissh `prace_service -i -s anselm`
 
 #### Access From Public Internet:
 
-It is recommended to use the single DNS name **name-cluster**.it4i.cz which is distributed between the four login nodes. If needed, user can login directly to one of the login nodes. The addresses are:
+It is recommended to use the single DNS name **name-cluster**.it4i.cz which is distributed between the four login nodes. If needed, the user can login directly to one of the login nodes. The addresses are:
 
-For Salomon cluster:
+Salomon cluster:
 
 | Login address                | Port | Protocol | Login node                       |
 | ---------------------------- | ---- | -------- | -------------------------------- |
@@ -106,7 +106,7 @@ For Salomon cluster:
 $ gsissh -p 2222 salomon.it4i.cz
 ```
 
-For Anselm cluster:
+Anselm cluster:
 
 | Login address         | Port | Protocol | Login node       |
 | --------------------- | ---- | -------- | ---------------- |
@@ -128,8 +128,7 @@ $ gsissh `prace_service -e -s salomon`
 $ gsissh `prace_service -e -s anselm`
 ```
 
-Although the preferred and recommended file transfer mechanism is [using GridFTP][5], the GSI SSH
-implementation supports also SCP, so for small files transfer gsiscp can be used:
+Although the preferred and recommended file transfer mechanism is [using GridFTP][5], the GSI SSH implementation also supports SCP, so for small files transfer, gsiscp can be used:
 
 ```console
 $ gsiscp -P 2222 _LOCAL_PATH_TO_YOUR_FILE_ salomon.it4i.cz:_SALOMON_PATH_TO_YOUR_FILE_
@@ -147,9 +146,9 @@ $ gsiscp -P 2222 anselm-prace.it4i.cz:_ANSELM_PATH_TO_YOUR_FILE_ _LOCAL_PATH_TO_
 
 ### Access to X11 Applications (VNC)
 
-If the user needs to run X11 based graphical application and does not have a X11 server, the applications can be run using VNC service. If the user is using regular SSH based access, see the [section in general documentation][6].
+If the user needs to run X11 based graphical application and does not have a X11 server, the applications can be run using VNC service. If the user is using a regular SSH based access, see this [section in general documentation][6].
 
-If the user uses GSI SSH based access, then the procedure is similar to the [SSH based access][6], only the port forwarding must be done using GSI SSH:
+If the user uses a GSI SSH based access, then the procedure is similar to the [SSH based access][6], only the port forwarding must be done using GSI SSH:
 
 ```console
 $ gsissh -p 2222 salomon.it4i.cz -L 5961:localhost:5961
@@ -157,19 +156,19 @@ $ gsissh -p 2222 salomon.it4i.cz -L 5961:localhost:5961
 
 ### Access With SSH
 
-After successful obtainment of login credentials for the local IT4Innovations account, the PRACE users can access the cluster as regular users using SSH. For more information see [the section in general documentation[9].
+After they successfully obtain the login credentials for the local IT4Innovations account, the PRACE users can access the cluster as regular users using SSH. For more information, see this [section in general documentation][9].
 
 ## File Transfers
 
-PRACE users can use the same transfer mechanisms as regular users (if they've undergone the full registration procedure). For information about this, see [the section in the general documentation][9].
+PRACE users can use the same transfer mechanisms as regular users (if they have undergone the full registration procedure). For more information, see the [Accessing the Clusters][9] section.
 
-Apart from the standard mechanisms, for PRACE users to transfer data to/from Salomon cluster, a GridFTP server running Globus Toolkit GridFTP service is available. The service is available from public Internet as well as from the internal PRACE network (accessible only from other PRACE partners).
+Apart from the standard mechanisms, for PRACE users to transfer data to/from the Salomon cluster, a GridFTP server running the Globus Toolkit GridFTP service is available. The service is available from public Internet as well as from the internal PRACE network (accessible only from other PRACE partners).
 
-There's one control server and three backend servers for striping and/or backup in case one of them would fail.
+There is one control server and three backend servers for striping and/or backup in case one of them would fail.
 
 ### Access From PRACE Network
 
-For Salomon cluster:
+Salomon cluster:
 
 | Login address                 | Port | Node role                   |
 | ----------------------------- | ---- | --------------------------- |
@@ -184,7 +183,7 @@ Copy files **to** Salomon by running the following commands on your local machin
 $ globus-url-copy file://_LOCAL_PATH_TO_YOUR_FILE_ gsiftp://gridftp-prace.salomon.it4i.cz:2812/home/prace/_YOUR_ACCOUNT_ON_SALOMON_/_PATH_TO_YOUR_FILE_
 ```
 
-For Anselm cluster:
+Anselm cluster:
 
 | Login address                | Port | Node role                   |
 | ---------------------------- | ---- | --------------------------- |
@@ -221,7 +220,7 @@ Copy files **from** Anselm:
 $ globus-url-copy gsiftp://gridftp-prace.anselm.it4i.cz:2812/home/prace/_YOUR_ACCOUNT_ON_ANSELM_/_PATH_TO_YOUR_FILE_ file://_LOCAL_PATH_TO_YOUR_FILE_
 ```
 
-Or by using prace_service script:
+Or by using the prace_service script:
 
 ```console
 $ globus-url-copy gsiftp://`prace_service -i -f salomon`/home/prace/_YOUR_ACCOUNT_ON_SALOMON_/_PATH_TO_YOUR_FILE_ file://_LOCAL_PATH_TO_YOUR_FILE_
@@ -233,7 +232,7 @@ $ globus-url-copy gsiftp://`prace_service -i -f anselm`/home/prace/_YOUR_ACCOUNT
 
 ### Access From Public Internet
 
-For Salomon cluster:
+Salomon cluster:
 
 | Login address           | Port | Node role                   |
 | ----------------------- | ---- | --------------------------- |
@@ -248,7 +247,7 @@ Copy files **to** Salomon by running the following commands on your local machin
 $ globus-url-copy file://_LOCAL_PATH_TO_YOUR_FILE_ gsiftp://gridftp.salomon.it4i.cz:2812/home/prace/_YOUR_ACCOUNT_ON_SALOMON_/_PATH_TO_YOUR_FILE_
 ```
 
-For Anselm cluster:
+Anselm cluster:
 
 | Login address          | Port | Node role                   |
 | ---------------------- | ---- | --------------------------- |
@@ -263,7 +262,7 @@ Copy files **to** Anselm by running the following commands on your local machine
 $ globus-url-copy file://_LOCAL_PATH_TO_YOUR_FILE_ gsiftp://gridftp.anselm.it4i.cz:2812/home/prace/_YOUR_ACCOUNT_ON_ANSELM_/_PATH_TO_YOUR_FILE_
 ```
 
-Or by using prace_service script:
+Or by using the prace_service script:
 
 ```console
 $ globus-url-copy file://_LOCAL_PATH_TO_YOUR_FILE_ gsiftp://`prace_service -e -f salomon`/home/prace/_YOUR_ACCOUNT_ON_SALOMON_/_PATH_TO_YOUR_FILE_
@@ -285,7 +284,7 @@ Copy files **from** Anselm:
 $ globus-url-copy gsiftp://gridftp.anselm.it4i.cz:2812/home/prace/_YOUR_ACCOUNT_ON_ANSELM_/_PATH_TO_YOUR_FILE_ file://_LOCAL_PATH_TO_YOUR_FILE_
 ```
 
-Or by using prace_service script:
+Or by using the prace_service script:
 
 ```console
 $ globus-url-copy gsiftp://`prace_service -e -f salomon`/home/prace/_YOUR_ACCOUNT_ON_SALOMON_/_PATH_TO_YOUR_FILE_ file://_LOCAL_PATH_TO_YOUR_FILE_
@@ -295,7 +294,7 @@ $ globus-url-copy gsiftp://`prace_service -e -f salomon`/home/prace/_YOUR_ACCOUN
 $ globus-url-copy gsiftp://`prace_service -e -f anselm`/home/prace/_YOUR_ACCOUNT_ON_ANSELM_/_PATH_TO_YOUR_FILE_ file://_LOCAL_PATH_TO_YOUR_FILE_
 ```
 
-Generally both shared file systems are available through GridFTP:
+Generally, both shared file systems are available through GridFTP:
 
 | File system mount point | Filesystem | Comment                                                        |
 | ----------------------- | ---------- | -------------------------------------------------------------- |
@@ -305,7 +304,7 @@ Generally both shared file systems are available through GridFTP:
 More information about the shared file systems is available [for Salomon here][10] and [for Anselm here][11].
 
 !!! hint
-    `prace` directory is used for PRACE users on the SCRATCH file system.
+    The `prace` directory is used for PRACE users on the SCRATCH file system.
 
 Only Salomon cluster /scratch:
 
@@ -316,15 +315,15 @@ Only Salomon cluster /scratch:
 
 ## Usage of the Cluster
 
-There are some limitations for PRACE user when using the cluster. By default PRACE users aren't allowed to access special queues in the PBS Pro to have high priority or exclusive access to some special equipment like accelerated nodes and high memory (fat) nodes. There may be also restrictions obtaining a working license for the commercial software installed on the cluster, mostly because of the license agreement or because of insufficient amount of licenses.
+There are some limitations for PRACE users when using the cluster. By default, PRACE users are not allowed to access special queues in the PBS Pro to have high priority or exclusive access to some special equipment like accelerated nodes and high memory (fat) nodes. There may also be restrictions on obtaining a working license for the commercial software installed on the cluster, mostly because of the license agreement or because of insufficient amount of licenses.
 
-For production runs always use scratch file systems. The available file systems are described [for Salomon here][10] and [for Anselm here][11].
+For production runs, always use scratch file systems. The available file systems are described [for Salomon here][10] and [for Anselm here][11].
 
 ### Software, Modules and PRACE Common Production Environment
 
-All system wide installed software on the cluster is made available to the users via the modules. The information about the environment and modules usage is in this [section of general documentation][12].
+All system-wide installed software on the cluster is made available to the users via the modules. For more information about the environment and modules usage, see the [Environment and Modules][12] section.
 
-PRACE users can use the "prace" module to use the [PRACE Common Production Environment][h].
+PRACE users can use the "prace" module for [PRACE Common Production Environment][h].
 
 ```console
 $ ml prace
@@ -332,11 +331,11 @@ $ ml prace
 
 ### Resource Allocation and Job Execution
 
-General information about the resource allocation, job queuing and job execution is in this [section of general documentation][13].
+For general information about the resource allocation, job queuing, and job execution, see [Resources Allocation Policy][13].
 
 For PRACE users, the default production run queue is "qprace". PRACE users can also use two other queues "qexp" and "qfree".
 
-For Salomon:
+Salomon:
 
 | queue                         | Active project | Project resources | Nodes                      | priority | authorization | walltime  |
 | ----------------------------- | -------------- | ----------------- | -------------------------- | -------- | ------------- | --------- |
@@ -344,7 +343,7 @@ For Salomon:
 | **qprace** Production queue   | yes            | >0                | 1006 nodes, max 86 per job | 0        | no            | 24 / 48 h |
 | **qfree** Free resource queue | yes            | none required     | 752 nodes, max 86 per job  | -1024    | no            | 12 / 12 h |
 
-For Anselm:
+Anselm:
 
 | queue                         | Active project | Project resources | Nodes               | priority | authorization | walltime  |
 | ----------------------------- | -------------- | ----------------- | ------------------- | -------- | ------------- | --------- |
@@ -352,15 +351,15 @@ For Anselm:
 | **qprace** Production queue   | yes            | > 0               | 178 w/o accelerator | medium   | no            | 24 / 48 h |
 | **qfree** Free resource queue | yes            | none required     | 178 w/o accelerator | very low | no            | 12 / 12 h |
 
-**qprace**, the PRACE This queue is intended for normal production runs. It is required that active project with nonzero remaining resources is specified to enter the qprace. The queue runs with medium priority and no special authorization is required to use it. The maximum runtime in qprace is 48 hours. If the job needs longer time, it must use checkpoint/restart functionality.
+**qprace**, the PRACE queue is intended for normal production runs. It is required that active project with nonzero remaining resources is specified to enter the qprace. The queue runs with medium priority and no special authorization is required to use it. The maximum runtime in qprace is 48 hours. If the job needs longer time, it must use the checkpoint/restart functionality.
 
 ### Accounting & Quota
 
-The resources that are currently subject to accounting are the core hours. The core hours are accounted on the wall clock basis. The accounting runs whenever the computational cores are allocated or blocked via the PBS Pro workload manager (the qsub command), regardless of whether the cores are actually used for any calculation. See [example in the general documentation][13].
+The resources that are currently subject to accounting are the core hours. The core hours are accounted on the wall clock basis. The accounting runs whenever the computational cores are allocated or blocked via the PBS Pro workload manager (the qsub command), regardless of whether the cores are actually used for any calculation. See the [example in the general documentation][13].
 
 PRACE users should check their project accounting using the [PRACE Accounting Tool (DART)][i].
 
-Users who have undergone the full local registration procedure (including signing the IT4Innovations Acceptable Use Policy) and who have received local password may check at any time, how many core-hours have been consumed by themselves and their projects using the command "it4ifree". You need to know your user password to use the command and that the displayed core hours are "system core hours" which differ from PRACE "standardized core hours".
+Users who have undergone the full local registration procedure (including signing the IT4Innovations Acceptable Use Policy) and who have received a local password may check at any time, how many core-hours they and their projects have consumed using the command "it4ifree". Note that you need to know your user password to use the command and that the displayed core hours are "system core hours" which differ from PRACE "standardized core hours".
 
 !!! note
     The **it4ifree** command is a part of it4i.portal.clients package, [located here][pypi].
@@ -388,7 +387,7 @@ WCH   =    Wall-clock Core Hour
 NCH   =    Normalized Core Hour
 ```
 
-By default file system quota is applied. To check the current status of the quota (separate for HOME and SCRATCH) use
+By default, a file system quota is applied. To check the current status of the quota (separate for HOME and SCRATCH), use:
 
 ```console
 $ quota