Last modified by valad on 2024/05/28 10:16

From version 1.1
edited by Thomas Coelho
on 2023/02/27 14:52
Change comment: There is no comment for this version
To version 2.1
edited by Thomas Coelho
on 2023/02/27 14:52
Change comment: There is no comment for this version

Summary

Details

Page properties
Syntax
... ... @@ -1,1 +1,1 @@
1 -MediaWiki 1.6
1 +XWiki 2.1
Content
... ... @@ -1,19 +1,19 @@
1 -We only support openMPI, currently the best MPI implementation. The widely used MPICH2 implementation, is difficult to maintain and Debian/Ubuntu packages are not available. The lamMPI implementation has stopped development in favour to openMPI.
1 +We only support openMPI, currently the best MPI implementation. The widely used MPICH2 implementation, is difficult to maintain and Debian/Ubuntu packages are not available. The lamMPI implementation has stopped development in favour to openMPI.
2 2  
3 -== Compiling ==
3 +== Compiling ==
4 4  
5 -To compile MPI programms your are recommended to use the commands mpicc, mpif77, mpif90. By default this wrapper refer to the GNU Compilers gcc and gfortran. To use the wrapper with the Intel Compilers you have to define some environment variables.
5 +To compile MPI programms your are recommended to use the commands mpicc, mpif77, mpif90. By default this wrapper refer to the GNU Compilers gcc and gfortran. To use the wrapper with the Intel Compilers you have to define some environment variables.
6 6  
7 -Bash users can add the following lines to their .bashrc:
8 -<pre>export OMPI_FC='ifort'
7 +Bash users can add the following lines to their .bashrc:
8 +
9 +{{{export OMPI_FC='ifort'
9 9  export OMPI_F77='ifort'
10 -export OMPI_CC='icc'
11 +export OMPI_CC='icc'}}}
11 11  
12 -</pre>
13 13  
14 -Especially if you use icc, it will be better if you use the recent [[Intel_Compiler_Temp|Intel Compiler 11.0]].
15 15  
15 +Especially if you use icc, it will be better if you use the recent [[Intel Compiler 11.0>>Intel_Compiler_Temp]].
16 +
16 16  == Infiniband ==
17 17  
18 18  The 'itp'-, 'itp-big'-, 'dfg-xeon'-, 'iboga'-, 'dreama'- and 'barcelona'-nodes have Infiniband Network. It is used by default when using our openMPI installation. Infiniband provides high bandwith with low latency. It can transport 20 GB/s with latency of 4 µs. In comparison, normal Gigabit Ethernet provides 1 GB/s with at least 30 µs latency. Latency is the time a data packet needs to travel from the source to its target. This is the main drawback when using Gigibit Ethernet for MPI communication.
19 -