From 9f944be5153e050227ba55fae05c7d1c6716086c Mon Sep 17 00:00:00 2001
From: Branislav Jansik <branislav.jansik@vsb.cz>
Date: Fri, 22 Mar 2019 18:38:44 +0100
Subject: [PATCH] Update introduction.md

---
 docs.it4i/dgx2/introduction.md | 19 +++++++++++++------
 1 file changed, 13 insertions(+), 6 deletions(-)

diff --git a/docs.it4i/dgx2/introduction.md b/docs.it4i/dgx2/introduction.md
index 1d7b43054..22e46cc68 100644
--- a/docs.it4i/dgx2/introduction.md
+++ b/docs.it4i/dgx2/introduction.md
@@ -1,8 +1,10 @@
 # NVIDIA DGX-2
 
-[DGX-2][a] builds upon [DGX-1][b] in several ways. Introduces NVIDIA’s new NVSwitch, enabling 300 GB/s chip-to-chip communication at 12 times the speed of PCIe.
+The [DGX-2][a] introduces NVIDIA’s new NVSwitch, enabling 300 GB/s chip-to-chip communication at 12 times the speed of PCIe.
 
-With NVLink2, enables sixteen GPUs to be grouped together in a single system, for a total bandwidth going beyond 14 TB/s. Pair of Xeon CPUs, 1.5 TB of memory, and 30 TB of NVMe storage, and we get a system that consumes 10 kW, weighs 163.29 kg, but offers easily double the performance of the DGX-1.
+With NVLink2, it enables sixteen Nvidia V100-SXM3 GPUs in a single system, for a total bandwidth going beyond 14 TB/s. 
+Featuring pair of Xeon 8168 CPUs, 1.5 TB of memory, and 30 TB of NVMe storage,
+we get a system that consumes 10 kW, weighs 163.29 kg, but offers perfomance in excess of 130TF.
 
 NVIDIA likes to tout that this means it offers a total of ~2 PFLOPs of compute performance in a single system, when using the tensor cores.
 
@@ -28,17 +30,22 @@ NVIDIA likes to tout that this means it offers a total of ~2 PFLOPs of compute p
 
 AlexNET, the network that 'started' the latest machine learning revolution, now takes 18 minutes
 
-The topology of the DGX-2 means that all 16 GPUs are able to pool their memory into a unified memory space, though with the usual tradeoffs involved if going off-chip.
+The topology of the DGX-2 means that all 16 GPUs are able to pool their memory into a unified memory space, 
+though with the usual tradeoffs involved if going off-chip.
 
-Not unlike the Tesla V100 memory capacity increase then, one of NVIDIA’s goals here is to build a system that can keep in-memory workloads that would be too large for an 8 GPU cluster. Providing one such example, NVIDIA is saying that the DGX-2 is able to complete the training process for FAIRSEQ – a neural network model for language translation – 10x faster than a DGX-1 system, bringing it down to less than two days total rather than 15.
+The DGX-2 is able to complete the training process 
+for FAIRSEQ – a neural network model for language translation – 10x faster than a DGX-1 system,
+bringing it down to less than two days total rather than 15.
 
 ![](../img/dgx3.png)
 
-Otherwise, similar to its DGX-1 counterpart, the DGX-2 is designed to be a powerful server in its own right. On the storage side the DGX-2 comes with 30TB of NVMe-based solid state storage. And for clustering or further inter-system communications, it also offers InfiniBand and 100GigE connectivity, up to eight of them.
+The DGX-2 is designed to be a powerful server in its own right. 
+On the storage side the DGX-2 comes with 30TB of NVMe-based solid state storage. 
+For clustering or further inter-system communications, it also offers InfiniBand and 100GigE connectivity, up to eight of them.
 
 ![](../img/dgx4.png)
 
 The new NVSwitches means that the PCIe lanes of the CPUs can be redirected elsewhere, most notably towards storage and networking connectivity.
 
 [a]: https://www.nvidia.com/content/dam/en-zz/es_em/Solutions/Data-Center/dgx-2/nvidia-dgx-2-datasheet.pdf
-[b]: https://www.nvidia.com/content/dam/en-zz/Solutions/Data-Center/dgx-1/dgx-1-ai-supercomputer-datasheet-v4.pdf
+
-- 
GitLab