Newer
Older
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
Compilers
=========
Available compilers, including GNU, INTEL and UPC compilers
Currently there are several compilers for different programming
languages available on the Anselm cluster:
- C/C++
- Fortran 77/90/95
- Unified Parallel C
- Java
- nVidia CUDA
The C/C++ and Fortran compilers are divided into two main groups GNU and
Intel.
Intel Compilers
---------------
For information about the usage of Intel Compilers and other Intel
products, please read the [Intel Parallel
studio](intel-suite.html) page.
GNU C/C++ and Fortran Compilers
-------------------------------
For compatibility reasons there are still available the original (old
4.4.6-4) versions of GNU compilers as part of the OS. These are
accessible in the search path by default.
It is strongly recommended to use the up to date version (4.8.1) which
comes with the module gcc:
$ module load gcc
$ gcc -v
$ g++ -v
$ gfortran -v
With the module loaded two environment variables are predefined. One for
maximum optimizations on the Anselm cluster architecture, and the other
for debugging purposes:
$ echo $OPTFLAGS
-O3 -march=corei7-avx
$ echo $DEBUGFLAGS
-O0 -g
For more informations about the possibilities of the compilers, please
see the man pages.
Unified Parallel C
------------------
UPC is supported by two compiler/runtime implementations:
- GNU - SMP/multi-threading support only
- Berkley - multi-node support as well as SMP/multi-threading support
### GNU UPC Compiler
To use the GNU UPC compiler and run the compiled binaries use the module
gupc
$ module add gupc
$ gupc -v
$ g++ -v
Simple program to test the compiler
$ cat count.upc
/* hello.upc - a simple UPC example */
#include <upc.h>
#include <stdio.h>
int main() {
if (MYTHREAD == 0) {
printf("Welcome to GNU UPC!!!n");
}
upc_barrier;
printf(" - Hello from thread %in", MYTHREAD);
return 0;
}
To compile the example use
$ gupc -o count.upc.x count.upc
To run the example with 5 threads issue
$ ./count.upc.x -fupc-threads-5
For more informations see the man pages.
### Berkley UPC Compiler
To use the Berkley UPC compiler and runtime environment to run the
binaries use the module bupc
$ module add bupc
$ upcc -version
As default UPC network the "smp" is used. This is very quick and easy
way for testing/debugging, but limited to one node only.
For production runs, it is recommended to use the native Infiband
implementation of UPC network "ibv". For testing/debugging using
multiple nodes, the "mpi" UPC network is recommended. Please note, that
the selection of the network is done at the compile time** and not at
runtime (as expected)!
Example UPC code:
$ cat hello.upc
/* hello.upc - a simple UPC example */
#include <upc.h>
#include <stdio.h>
int main() {
if (MYTHREAD == 0) {
printf("Welcome to Berkeley UPC!!!n");
}
upc_barrier;
printf(" - Hello from thread %in", MYTHREAD);
return 0;
}
To compile the example with the "ibv" UPC network use
$ upcc -network=ibv -o hello.upc.x hello.upc
To run the example with 5 threads issue
$ upcrun -n 5 ./hello.upc.x
To run the example on two compute nodes using all 32 cores, with 32
threads, issue
$ qsub -I -q qprod -A PROJECT_ID -l select=2:ncpus=16
$ module add bupc
$ upcrun -n 32 ./hello.upc.x
For more informations see the man pages.
Java
----
For information how to use Java (runtime and/or compiler), please read
the [Java page](java.html).
nVidia CUDA
-----------
For information how to work with nVidia CUDA, please read the [nVidia
CUDA page](nvidia-cuda.html).