This FAQ is for Open MPI v4.x and earlier.
If you are looking for documentation for Open MPI v5.x and later, please visit docs.open-mpi.org.
Table of contents:
- How do I run jobs under Slurm?
- Does Open MPI support "srun -n X my_mpi_application"?
- I use Slurm on a cluster with the OpenFabrics network stack. Do I need to do anything special?
- My job fails / performs poorly when using mpirun under Slurm 20.11
1. How do I run jobs under Slurm? |
The short answer is yes, provided you configured OMPI
--with-slurm . You can use mpirun as normal, or directly launch
your application using srun if OMPI is configured per
this FAQ entry.
The longer answer is that Open MPI supports launching parallel jobs in
all three methods that Slurm supports (you can find more info about
Slurm specific recommendations on the SchedMD web page:
- Launching via "
salloc ... "
- Launching via "
sbatch ... "
- Launching via "
srun -n X my_mpi_application "
Specifically, you can launch Open MPI's mpirun in an interactive
Slurm allocation (via the salloc command) or you can submit a
script to Slurm (via the sbatch command), or you can "directly"
launch MPI executables via srun .
Open MPI automatically obtains both the list of hosts and how many
processes to start on each host from Slurm directly. Hence, it is
unnecessary to specify the --hostfile , --host , or -np options to
mpirun . Open MPI will also use Slurm-native mechanisms to launch
and kill processes (rsh and/or ssh are not required).
For example:
1
2
3
4
5
6
7
| # Allocate a Slurm job with 4 nodes
shell$ salloc -N 4 sh
# Now run an Open MPI job on all the nodes allocated by Slurm
# (Note that you need to specify -np for the 1.0 and 1.1 series;
# the -np value is inferred directly from Slurm starting with the
# v1.2 series)
shell$ mpirun my_mpi_application |
This will run the 4 MPI processes on the nodes that were allocated by
Slurm. Equivalently, you can do this:
1
2
| # Allocate a Slurm job with 4 nodes and run your MPI application in it
shell$ salloc -N 4 mpirun my_mpi_aplication |
Or, if submitting a script:
1
2
3
4
5
6
| shell$ cat my_script.sh
#!/bin/sh
mpirun my_mpi_application
shell$ sbatch -N 4 my_script.sh
srun: jobid 1234 submitted
shell$ |
2. Does Open MPI support "srun -n X my_mpi_application"? |
Yes, if you have configured OMPI --with-pmi=foo , where foo is
the path to the directory where pmi.h/pmi2.h is located. Slurm (> 2.6,
> 14.03) installs PMI-2 support by default.
Older versions of Slurm install PMI-1 by default. If you desire PMI-2,
Slurm requires that you manually install that support. When the
--with-pmi option is given, OMPI will automatically determine if PMI-2
support was built and use it in place of PMI-1.
3. I use Slurm on a cluster with the OpenFabrics network stack. Do I need to do anything special? |
Yes. You need to ensure that Slurm sets up the locked memory
limits properly. Be sure to see this FAQ entry about
locked memory and this FAQ entry for
references about Slurm.
4. My job fails / performs poorly when using mpirun under Slurm 20.11 |
There were some changes in Slurm behavior that were introduced
in Slurm 20.11.0 and subsequently reverted out in Slurm 20.11.3.
SchedMD (the makers of Slurm) strongly suggest that all Open MPI users
avoid using Slurm versions 20.11.0 through 20.11.2.
Indeed, you will likely run into problems using just about any version
of Open MPI these problematic Slurm releases. Please either downgrade
to an older version or upgrade to a newer version of Slurm.
|