Approvals: 0/1
This is an old revision of the document!
OpenFOAM and local temp space on VSC-1
a simple jobscript for using the /work/tmp directory on VSC1 for openFOAM:
OpenFOAM compile scripts
INTEL icc and INTELMPI on VSC-1
in July 2013 VSC-1 has been updated to Scientific Linux 6.4 (identical with VSC-2). Also the latest Infiniband Software only includes Intel MPI. For installing OpenFoam on VSC-1 please use the installation Scripts from VSC-2
(old): INTEL icc and Qlogic MPI on VSC-1
Set the MPI Version to 'qlogicmpi_intel-0.1.0', logout and login again.
Download the source packages for OpenFoam and Thirdparties. Place them at the same directory as the scripts which are provided below.
(old) Version 1.7.1
Patched File: turbulentheatfluxtemperaturefvpatchscalarfield.c
Place at same directory as the script is executed.
(old) Version 2.1.1
INTEL icc and Intel MPI on VSC-2
Set the MPI Version to 'intel_mpi_intel64-4.0.3.008' (1.7.1, 2.1.1) or 'impi_intel-4.1.0.024' (2.2.0), logout and login again.
Download the source packages for OpenFoam and Thirdparties. Place them at the same directory as the scripts which are provided below.
Version 1.7.1
Patched File: turbulentheatfluxtemperaturefvpatchscalarfield.c
Place at same directory as the script is executed.
Version 2.1.1
Version 2.2.0
Version 2.2.2
Version 2.3.x
version using the latest git repository of openfoam:
intel mpi 4.1.0.024 installopenfoam_2.3.x_vsc2_git_intel_mpi.4.1.0.024.sh
intel mpi 4.1.3.045 installopenfoam_2.3.x_vsc2_git_intel_mpi.4.1.3.045.sh
Version foam-ext-3.0
script may needed to be run twice, because git clone some how terminates the script …
Scaling of OpenFOAM
Performance tests were done using different setups, depending on the used cluster:
VSC-1
- Intel C-compiler 11.1
- Qlogic MPI 0.1.0
VSC-2
- Intel C-compiler 12.1.3
- Intel MPI 4.0.3.008 / impi_intel-4.1.0.024
Compared to the gcc + openmpi Version a speed up of about 1/3 was found in case of 32 cores on VSC-1
Note: the runs denoted “VSC1 qlogic scaled” are the same as “VSC1 qlogic” but the runtimes were multiplied by a factor of 1.33 which is the nominal factor in clockspeed between VSC-1 and VSC-2
Cavity Benchmark for version 2.2.0
cores | VSC-2 intel.4.1.0.024 mpich16 | VSC-2 intel.4.1.0.024 mpich8 |
---|---|---|
16 | 4471 | 4949 |
32 | 4753 | 2202 |
64 | 1832 | 809 |
128 | 508 | 279 |
256 | 234 | 182 |
Cavity Benchmark for version 2.1.1
real times for benchmark in seconds:
- mpich16: 16 cores per node were used for calculation
- mpich8: 8 cores per node were used for calculation
cores | VSC-1 qlogic | VSC-2 intel.4.0.3.008 mpich16 | VSC-2 intel.4.0.3.008 mpich8 | VSC-1 gcc + openmpi |
---|---|---|---|---|
16 | 2843 | 4426 | 3083 | |
32 | 1282 | 2176 | 1504 | 1818 |
64 | 545 | 937 | 663 | |
128 | 229 | 360 | 270 | |
256 | 134 | 207 | 173 |
Cavity benchmark for version 1.7.1
real times for benchmark in seconds:
- mpich16: 16 cores per node were used for calculation
cores | VSC-1 qlogic | VSC-2 intel.4.0.3.008 mpich16 |
---|---|---|
16 | 2826 | 4473 |
32 | 1260 | 2086 |
64 | 517 | 887 |
128 | 224 | 333 |
256 | - | 225 |
FoamPro
was used; ask admins