Differences between revisions 8 and 9
Revision 8 as of 2007-02-20 18:07:01
Size: 2733
Editor: GötzWaschk
Comment:
Revision 9 as of 2007-05-15 10:53:42
Size: 2116
Editor: GötzWaschk
Comment: Änderungen für SL5
Deletions are marked like this. Additions are marked like this.
Line 6: Line 6:
Since the upgrade to SL5, both clusters use the openmpi implementation of the MPI standard.
Line 7: Line 8:
=== Infiniband ===
Applications for the cluster must be compiled on a 64 bit machine, at the moment, this is lx64 only.
There are MPI versions for the GCC, Intel and PGI compilers installed:
There are MPI versions for the GCC, Intel and PGI compilers installed:
Line 10: Line 10:
/usr/local/ibgd/mpi/osu/gcc/mvapich-0.9.5/bin/mpicc /opt/openmpi/gcc/bin/mpicc
Line 12: Line 12:
/usr/local/ibgd/mpi/osu/intel/mvapich-0.9.5/bin/mpicc /opt/openmpi/intel/bin/mpicc
Line 14: Line 14:
/usr/local/ibgd/mpi/osu/pgi/mvapich-0.9.5/bin/mpicc
/opt/openmpi/pgi/bin/mpicc
Line 19: Line 18:

=== Infiniband ===
Applications for the cluster must be compiled on a 64 bit SL5 machine, at the moment, this is sl5-64.ifh.de only.
Line 20: Line 23:
Applications for this cluster can be compiled on the pub.ifh.de machines. There are MPI versions for the GCC, Intel and PGI compilers installed:

/opt/mpich/gcc/bin/mpicc

/opt/mpich/intel/bin/mpicc

/opt/mpich/pgi/bin/mpicc

Compilers for C++ and Fortran are available as well.
Applications for the cluster must be compiled on a 32 bit SL5 machine, at the moment, this is sl5.ifh.de only.
Line 40: Line 35:
Be sure to call the right mpirun version. On the Infiniband cluster use: Be sure to call the right mpirun version for your compiler. If you application was compiled with GCC, use
Line 42: Line 37:
/usr/local/ibgd/mpi/osu/gcc/mvapich-0.9.5/bin/mpirun -np $NSLOTS -machinefile $TMPDIR/machines yourapp


On the Myrinet cluster use:

/opt/mpich/gcc/bin/mpirun -np $NSLOTS -machinefile $TMPDIR/machines yourapp
/opt/openmpi/gcc/bin/mpirun -np $NSLOTS yourapp
Line 58: Line 48:
If your application is using threads, it is recommended to set the value of h_stack (by default the same as h_vmem) to a sane value, e.g. 10M.
Line 62: Line 54:
Be aware that the batch system renews the AFS token, but only on the node that starts the first process (node 0). That's why you should access the AFS from that node. An example scenario looks like this:

 1. Copy data from AFS to node 0.
 1. Copy it with scp to the nodes that need it to the directory $TMPDIR, the machine names are in $TMPDIR/machines
 1. Run your MPI job.
 1. Copy the results with scp from the local discs to node 0.
 1. Copy the data from node 0 to AFS.
Line 72: Line 56:

Usage of the Linux Clusters at DESY Zeuthen

At Zeuthen, two clusters are available, one with 16 dual Opteron machines connected by Infiniband and one with 8 dual Xeons and Myrinet. They are integrated into the SGE batch system. The documentation in ["Batch System Usage"] applies to them.

Building Applications

Since the upgrade to SL5, both clusters use the openmpi implementation of the MPI standard.

There are MPI versions for the GCC, Intel and PGI compilers installed:

/opt/openmpi/gcc/bin/mpicc

/opt/openmpi/intel/bin/mpicc

/opt/openmpi/pgi/bin/mpicc

Compilers for C++ and Fortran are available as well.

Infiniband

Applications for the cluster must be compiled on a 64 bit SL5 machine, at the moment, this is sl5-64.ifh.de only.

Myrinet

Applications for the cluster must be compiled on a 32 bit SL5 machine, at the moment, this is sl5.ifh.de only.

Batch System Access

A job script designated for a parallel job needs to specify the parallel environment and the number of required CPUs. The parameter looks like this for the Infiniband cluster:

#$ -pe mpich-ppn2 4

On the Myrinet cluster, it is similar:

#$ -pe mpichgm-ppn2 4

Be sure to call the right mpirun version for your compiler. If you application was compiled with GCC, use

/opt/openmpi/gcc/bin/mpirun -np $NSLOTS yourapp

It is important to request the right limit for memory with the parameter h_vmem.

The Opteron machines have 3.3G of RAM and by default two jobs are executed on one node, so the maximal amount of memory is 1650M per process:

#$ -l h_vmem=1650M

The Xeons have 922.5M of RAM.

If your application is using threads, it is recommended to set the value of h_stack (by default the same as h_vmem) to a sane value, e.g. 10M.

AFS Access

The application binary must be available to all nodes, that's why it should be placed in an AFS directory.

Further documentation

[http://www-zeuthen.desy.de/technisches_seminar/texte/Technisches_Seminar_Waschk.pdf HPC-Clusters at DESY Zeuthen] , 11/22/06, technical seminar

Cluster (last edited 2023-04-28 09:56:09 by GötzWaschk)