[slurm-users] Multinode MPI job

Christopher Samuel chris at csamuel.org
Wed Mar 27 16:09:45 UTC 2019

On 3/27/19 8:39 AM, Mahmood Naderan wrote:

> mpirun pw.x -imos2.rlx.in  <http://mos2.rlx.in>

You will need to read the documentation for this:


Especially note both of these:

IMPORTANT: The ability to execute a single application across more than 
one job allocation does not work with all MPI implementations or Slurm 
MPI plugins. Slurm's ability to execute such an application can be 
disabled on the entire cluster by adding "disable_hetero_steps" to 
Slurm's SchedulerParameters configuration parameter.

IMPORTANT: While the srun command can be used to launch heterogeneous 
job steps, mpirun would require substantial modification to support 
heterogeneous applications. We are aware of no such mpirun development 
efforts at this time.

So at the very least you'll need to use srun, not mpirun and confirm 
that the MPI you are using supports this Slurm feature.

> Also, the partition names are weird. We have these entries:

Partitions are defined by the systems administrators, you'd need to 
speak with them about their reasoning for those.

All the best,
   Chris Samuel  :  http://www.csamuel.org/  :  Berkeley, CA, USA

More information about the slurm-users mailing list