Skip to content


The Message Passing Interface (MPI) implementation is an important part of an HPC Vega cluster. The vast majority of software that needs to communicate across nodes uses MPI. Although it is not the only means of communication between processes, it is the only one that we provide a standard implementation for that can be used both for compiling and running software. Other more specialized communication implementations are typically built with the software that utilizes them.

More information about MPI can be found at the following location:

There are several implementations of the MPI standard. The one that we use on the HPC Vega is Open MPI.

Open MPI on HPC Vega is available as a module. Search and display for the desired module:

$ module keyword openmpi/
$ module whatis openmpi/

# Check all options with `tab`
$ module load openmpi/<tab>

Loading the selected module (Default Open MPI version is

$ module load openmpi/

Once you have loaded the Open MPI module then the mpirun, mpiexec, mpicc, mpif90 commands will be available for running your MPI programs.

$ mpirun ./my_application

Single node jobs

Use any version on HPC Vega (eg. cvmfs modules mounted at /cvmfs available through modules or system Open MPI).

Multi node jobs

HPC Vega is using Infiniband, therefore for successful job execution correct environment must be setup.

The system Open MPI module (4.0.5) is already configured and loaded by default. btl disabled by default, therefore UCX takes correct interface (ib0).

Usage of srun:

Check supported API interfaces with command:

srun --mpi=list

We recommend usage of pmix (eg. GNU compiler). Example of SBATCH script:

#SBATCH --jobname=multinode-srun-test
#SBATCH --nodes=2
#SBATVH --ntasks=4
#SBATCH --ntasks-per-node=2
#SBATCH --time=00:15:00
#SBATCH --output %j.output
#SBATCH --partition=cpu

srun --mpi=pmix ./mpibench 

For Intel, use pmi2 and export environment variable I_MPI_PMI_LIBRARY.

export I_MPI_PMI_LIBRARY=/usr/lib64/slurm/
srun --mpi=pmi2 ./mpibench

Usage of mpirun:

Open MPI, version 3.x

$ mpirun -mca btl_tcp_if_include ib0 ./mpibench

Open MPI, version prior to 4.0.3

Enable ucx and disable btl. Example of SBATCH script:

#SBATCH --job-name=multinode-mpirun-test
#SBATCH --nodes=2
#SBATVH --ntasks=4
#SBATCH --ntasks-per-node=2
#SBATCH --time=00:15:00
#SBATCH --output %j.output
#SBATCH --partition=cpu

module load OpenMPI/4.0.3-GCC-9.3.0

mpirun -mca pml ucx -mca btl ^uct,tcp,openib,vader --bind-to core:overload-allowed ./mpibench

mpibench can be found at:

  • Source code: mpiBench
  • Compiled code: /ceph/hpc/software/mpibench

For more information for multithreaded, multicore and other jobs with Open MPI, Open MP, hybrid can be found at: LINK.