diff --git a/docs.it4i/barbora/compute-nodes.md b/docs.it4i/barbora/compute-nodes.md
index 9915e01c3166dbd9bf12279fe61a156f9e65c448..a1f4f53d5086dfeef61535406906a13d45222831 100644
--- a/docs.it4i/barbora/compute-nodes.md
+++ b/docs.it4i/barbora/compute-nodes.md
@@ -6,13 +6,13 @@ Barbora is a cluster of x86-64 Intel-based nodes built with the BullSequana Comp
 
 * 192 nodes
 * 6912 cores in total
-* 2x Intel Cascade Lake 6240, 18-core, 2.6 GHz processors per node
-* 192 GB DDR4 2933MT/s of physical memory per node (12x 16 GB)
+* 2x Intel Cascade Lake 6240, 18-core, 2.6GHz processors per node
+* 192GB DDR4 2933MT/s of physical memory per node (12x16GB)
 * BullSequana X1120 blade servers
-* 2995,2 GFLOP/s per compute node
-* 1x 1 GB Ethernet
+* 2995.2 GFLOP/s per compute node
+* 1x 1GB Ethernet
 * 1x HDR100 IB port
-* 3 computes nodes per X1120 blade server
+* 3 compute nodes per X1120 blade server
 * cn[1-192]
 
 ![](img/BullSequanaX1120.png)
@@ -21,13 +21,13 @@ Barbora is a cluster of x86-64 Intel-based nodes built with the BullSequana Comp
 
 * 8 nodes
 * 192 cores in total
-* two Intel Skylake Gold 6126, 12-core, 2.6 GHz processors per node
-* 192 GB DDR4 2933MT/s with ECC of physical memory per node (12x 16 GB)
+* two Intel Skylake Gold 6126, 12-core, 2.6GHz processors per node
+* 192 GB DDR4 2933MT/s with ECC of physical memory per node (12x16GB)
 * 4x GPU accelerator NVIDIA Tesla V100-SXM2 per node
 * Bullsequana X410-E5 NVLink-V blade servers
-* 1996,8 GFLOP/s per compute nodes
-* GPU-tp-GPU All-to-All NVLINK 2.0, GPU-Direct
-* 1 GB Ethernet
+* 1996.8 GFLOP/s per compute nodes
+* GPU-to-GPU All-to-All NVLINK 2.0, GPU-Direct
+* 1GB Ethernet
 * 2x HDR100 IB ports
 * cn[193-200]
 
@@ -37,8 +37,8 @@ Barbora is a cluster of x86-64 Intel-based nodes built with the BullSequana Comp
 
 * 1x BullSequana X808 server
 * 128 cores in total
-* 8 Intel Skylake 8153, 16-core, 2.0 GHz, 125W
-* 6144 GiB DDR4 2667MT/s of physical memory per node (92x 64 GB)
+* 8 Intel Skylake 8153, 16-core, 2.0GHz, 125W
+* 6144 GiB DDR4 2667MT/s of physical memory per node (92x64GB)
 * 2x HDR100 IB port
 * 8192 GFLOP/s
 * cn[201]
@@ -47,11 +47,11 @@ Barbora is a cluster of x86-64 Intel-based nodes built with the BullSequana Comp
 
 ## Compute Node Summary
 
-| Node type                    | Count | Range       | Memory  | Cores         | Queues                     |
-| ---------------------------- | ----- | ----------- | ------  | -----------   | -------------------------- |
-| Nodes without an accelerator | 192   | cn[1-192]   | 192GB   | 36 @ 2.6 GHz  | qexp, qprod, qlong, qfree  |
-| Nodes with a GPU accelerator | 8     | cn[193-200] | 192GB   | 24 @ 2.6 GHz  | qnvidia                    |
-| Fat compute nodes            | 1     | cn[201]     | 6144GiB | 128 @ 2.0 GHz | qfat                       |
+| Node type                    | Count | Range       | Memory  | Cores        |
+| ---------------------------- | ----- | ----------- | ------  | -----------  |
+| Nodes without an accelerator | 192   | cn[1-192]   | 192GB   | 36 @ 2.6GHz  |
+| Nodes with a GPU accelerator | 8     | cn[193-200] | 192GB   | 24 @ 2.6GHz  |
+| Fat compute nodes            | 1     | cn[201]     | 6144GiB | 128 @ 2.0GHz |
 
 ## Processor Architecture
 
@@ -116,23 +116,23 @@ Barbora is equipped with an [NVIDIA Tesla V100-SXM2][g] accelerator.
 
 ![](img/gpu-v100.png)
 
-|NVIDIA Tesla V100-SXM2||
-| --- | --- |
-| GPU Architecture | NVIDIA Volta |
-| NVIDIA Tensor| Cores: 640 |
-| NVIDIA CUDA® Cores | 5 120 |
-| Double-Precision Performance | 7.8 TFLOP/s |
-| Single-Precision Performance | 15.7 TFLOP/s |
-| Tensor Performance | 125 TFLOP/s |
-| GPU Memory | 16 GB HBM2 |
-| Memory Bandwidth | 900 GB/sec |
-| ECC | Yes |
-| Interconnect Bandwidth | 300 GB/sec |
-| System Interface | NVIDIA NVLink |
-| Form Factor | SXM2 |
-| Max Power Consumption | 300 W |
-| Thermal Solution | Passive |
-| Compute APIs | CUDA, DirectCompute,OpenCLTM, OpenACC |
+| NVIDIA Tesla V100-SXM2       |                                        |
+| ---------------------------- | -------------------------------------- |
+| GPU Architecture             | NVIDIA Volta                           |
+| NVIDIA Tensor Cores          | 640                                    |
+| NVIDIA CUDA® Cores           | 5120                                   |
+| Double-Precision Performance | 7.8TFLOP/s                             |
+| Single-Precision Performance | 15.7TFLOP/s                            |
+| Tensor Performance           | 125TFLOP/s                             |
+| GPU Memory                   | 16GB HBM2                              |
+| Memory Bandwidth             | 900GB/sec                              |
+| ECC                          | Yes                                    |
+| Interconnect Bandwidth       | 300GB/sec                              |
+| System Interface             | NVIDIA NVLink                          |
+| Form Factor                  | SXM2                                   |
+| Max Power Consumption        | 300W                                   |
+| Thermal Solution             | Passive                                |
+| Compute APIs                 | CUDA, DirectCompute, OpenCLTM, OpenACC |
 
 [a]: https://en.wikichip.org/wiki/intel/microarchitectures/skylake_(server)#Core
 [b]: https://en.wikichip.org/wiki/intel/microarchitectures/skylake_(client)#Pipeline