Skip to content
Snippets Groups Projects
Commit dfc7f760 authored by Pavel Jirásek's avatar Pavel Jirásek
Browse files

Merge branch 'content_revision' into 'master'

Content revision

See merge request !17
parents a3749016 389f85a2
No related branches found
No related tags found
No related merge requests found
......@@ -11,7 +11,7 @@ The Anselm cluster is accessed by SSH protocol via login nodes login1 and login2
|login1.anselm.it4i.cz|22|ssh|login1|
|login2.anselm.it4i.cz|22|ssh|login2|
The authentication is by the [private key](../../../get-started-with-it4innovations/accessing-the-clusters/shell-access-and-data-transfer/ssh-keys/)
The authentication is by the [private key](../get-started-with-it4innovations/accessing-the-clusters/shell-access-and-data-transfer/ssh-keys/)
!!! Note "Note"
Please verify SSH fingerprints during the first logon. They are identical on all login nodes:
......@@ -33,7 +33,7 @@ If you see warning message "UNPROTECTED PRIVATE KEY FILE!", use this command to
local $ chmod 600 /path/to/id_rsa
```
On **Windows**, use [PuTTY ssh client](../../../get-started-with-it4innovations/accessing-the-clusters/shell-access-and-data-transfer/putty/putty/).
On **Windows**, use [PuTTY ssh client](../get-started-with-it4innovations/accessing-the-clusters/shell-access-and-data-transfer/putty.md).
After logging in, you will see the command prompt:
......@@ -54,9 +54,8 @@ Last login: Tue Jul 9 15:57:38 2013 from your-host.example.com
Example to the cluster login:
<tty-player controls src=/src/anselm/login_anselm.ttyrec></tty-player>
>The environment is **not** shared between login nodes, except for [shared filesystems](../storage/storage/#section-1).
!!! Note "Note"
The environment is **not** shared between login nodes, except for [shared filesystems](../storage/storage/#shared-filesystems).
Data Transfer
-------------
......@@ -69,7 +68,7 @@ Data in and out of the system may be transferred by the [scp](http://en.wikipedi
|login2.anselm.it4i.cz|22|scp, sftp|
|dm1.anselm.it4i.cz|22|scp, sftp|
The authentication is by the [private key](../../../get-started-with-it4innovations/accessing-the-clusters/shell-access-and-data-transfer/ssh-keys/)
The authentication is by the [private key](../get-started-with-it4innovations/accessing-the-clusters/shell-access-and-data-transfer/ssh-keys/)
!!! Note "Note"
Data transfer rates up to **160MB/s** can be achieved with scp or sftp.
......@@ -86,8 +85,6 @@ On linux or Mac, use scp or sftp client to transfer the data to Anselm:
```bash
local $ scp -i /path/to/id_rsa my-local-file username@anselm.it4i.cz:directory/file
```
For example how to using scp command:
<tty-player controls src=/src/anselm/transfer_scp_anselm.ttyrec></tty-player>
```bash
local $ scp -i /path/to/id_rsa -r my-local-dir username@anselm.it4i.cz:directory
......@@ -98,8 +95,6 @@ or
```bash
local $ sftp -o IdentityFile=/path/to/id_rsa username@anselm.it4i.cz
```
For example how to using sftp command:
<tty-player controls src=/src/anselm/transfer_sftp_anselm.ttyrec></tty-player>
Very convenient way to transfer files in and out of the Anselm computer is via the fuse filesystem [sshfs](http://linux.die.net/man/1/sshfs)
......@@ -119,7 +114,7 @@ $ man sshfs
On Windows, use [WinSCP client](http://winscp.net/eng/download.php) to transfer the data. The [win-sshfs client](http://code.google.com/p/win-sshfs/) provides a way to mount the Anselm filesystems directly as an external disc.
More information about the shared file systems is available [here](../../storage/storage/).
More information about the shared file systems is available [here](../storage/storage/).
Connection restrictions
......@@ -170,8 +165,7 @@ Note: Port number 6000 is chosen as an example only. Pick any free port.
Remote port forwarding from compute nodes allows applications running on the compute nodes to access hosts outside Anselm Cluster.
First, establish the remote port forwarding form the login node, as [described
above](outgoing-connections.html#port-forwarding-from-login-nodes).
First, establish the remote port forwarding form the login node, as [described above](../anselm-cluster-documentation/shell-and-data-access/#port-forwarding-from-login-nodes).
Second, invoke port forwarding from the compute node to the login node. Insert following line into your jobscript or interactive shell
......@@ -196,7 +190,7 @@ local $ ssh -D 1080 localhost
On Windows, install and run the free, open source [Sock Puppet](http://sockspuppet.com/) server.
Once the proxy server is running, establish ssh port forwarding from Anselm to the proxy server, port 1080, exactly as [described above](outgoing-connections/#port-forwarding-from-login-nodes).
Once the proxy server is running, establish ssh port forwarding from Anselm to the proxy server, port 1080, exactly as [described above](../anselm-cluster-documentation/shell-and-data-access/#port-forwarding-from-login-nodes).
```bash
local $ ssh -R 6000:localhost:1080 anselm.it4i.cz
......
......@@ -20,7 +20,6 @@ The authentication is by the [private key](../get-started-with-it4innovations/ac
!!! Note "Note"
Please verify SSH fingerprints during the first logon. They are identical on all login nodes:
f6:28:98:e4:f9:b2:a6:8f:f2:f4:2d:0a:09:67:69:80 (DSA)
70:01:c9:9a:5d:88:91:c7:1b:c0:84:d1:fa:4e:83:5c (RSA)
......@@ -38,23 +37,17 @@ If you see warning message "UNPROTECTED PRIVATE KEY FILE!", use this command to
local $ chmod 600 /path/to/id_rsa
```
On **Windows**, use [PuTTY ssh client](../get-started-with-it4innovations/accessing-the-clusters/shell-access-and-data-transfer/putty/putty/).
On **Windows**, use [PuTTY ssh client](../get-started-with-it4innovations/accessing-the-clusters/shell-access-and-data-transfer/putty.md).
After logging in, you will see the command prompt:
<tty-player controls src=/src/salomon/login_salomon.ttyrec></tty-player>
!!! Note "Note"
The environment is **not** shared between login nodes, except for [shared filesystems](storage/storage/).
The environment is **not** shared between login nodes, except for [shared filesystems](../storage/storage/).
Data Transfer
-------------
Data in and out of the system may be transferred by the [scp](http://en.wikipedia.org/wiki/Secure_copy) and sftp protocols.
In case large volumes of data are transferred, use dedicated data mover nodes cedge[1-3].salomon.it4i.cz for increased performance.
HTML commented section #1 (removed cedge servers from the table)
|Address|Port|Protocol|
|---|---|
|salomon.it4i.cz|22|scp, sftp|
......@@ -65,17 +58,12 @@ HTML commented section #1 (removed cedge servers from the table)
The authentication is by the [private key](../get-started-with-it4innovations/accessing-the-clusters/shell-access-and-data-transfer/ssh-keys/)
HTML commented section #2 (ssh transfer performance data need to be verified)
On linux or Mac, use scp or sftp client to transfer the data to Salomon:
```bash
local $ scp -i /path/to/id_rsa my-local-file username@salomon.it4i.cz:directory/file
```
For example how to using scp command:
<tty-player controls src=/src/salomon/transfer_scp_salomon.ttyrec></tty-player>
```bash
local $ scp -i /path/to/id_rsa -r my-local-dir username@salomon.it4i.cz:directory
```
......@@ -86,9 +74,6 @@ or
local $ sftp -o IdentityFile=/path/to/id_rsa username@salomon.it4i.cz
```
For example how to using sftp command:
<tty-player controls src=/src/salomon/transfer_sftp_salomon.ttyrec></tty-player>
Very convenient way to transfer files in and out of the Salomon computer is via the fuse filesystem [sshfs](http://linux.die.net/man/1/sshfs)
```bash
......@@ -107,7 +92,7 @@ $ man sshfs
On Windows, use [WinSCP client](http://winscp.net/eng/download.php) to transfer the data. The [win-sshfs client](http://code.google.com/p/win-sshfs/) provides a way to mount the Salomon filesystems directly as an external disc.
More information about the shared file systems is available [here](storage/storage/).
More information about the shared file systems is available [here](../storage/storage/).
Connection restrictions
-----------------------
......@@ -157,7 +142,7 @@ Note: Port number 6000 is chosen as an example only. Pick any free port.
Remote port forwarding from compute nodes allows applications running on the compute nodes to access hosts outside Salomon Cluster.
First, establish the remote port forwarding form the login node, as [described above](outgoing-connections/#port-forwarding-from-login-nodes).
First, establish the remote port forwarding form the login node, as [described above](../salomon/shell-and-data-access/#port-forwarding-from-login-nodes).
Second, invoke port forwarding from the compute node to the login node. Insert following line into your jobscript or interactive shell
......@@ -182,10 +167,10 @@ local $ ssh -D 1080 localhost
On Windows, install and run the free, open source [Sock Puppet](http://sockspuppet.com/) server.
Once the proxy server is running, establish ssh port forwarding from Salomon to the proxy server, port 1080, exactly as [described above](outgoing-connections/#port-forwarding-from-login-nodes).
Once the proxy server is running, establish ssh port forwarding from Salomon to the proxy server, port 1080, exactly as [described above](../salomon/shell-and-data-access/#port-forwarding-from-login-nodes).
```bash
local $ ssh -R 6000:localhost:1080 salomon.it4i.cz
```
Now, configure the applications proxy settings to **localhost:6000**. Use port forwarding to access the [proxy server from compute nodes](outgoing-connections/#port-forwarding-from-compute-nodes) as well .
Now, configure the applications proxy settings to **localhost:6000**. Use port forwarding to access the [proxy server from compute nodes](../salomon/shell-and-data-access/#port-forwarding-from-compute-nodes) as well .
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment