[slurm-users] Segmentation fault when launching mpi jobs using Intel MPI

Bob Smith r3xj220 at gmail.com
Wed Feb 6 17:06:01 UTC 2019

Hello all,

I am having an issue submitting mpi jobs via sbatch using Intel MPI 2019
Update 1. The job ends with a segmentation fault immediately:

[user at head mpihello]$ cat mpihello-intelmpi.out.62
srun: error: node003: task 0: Segmentation fault
srun: error: node004: task 0: Segmentation fault
[mpiexec at node002] main
(../../../../../src/pm/i_hydra/mpiexec/mpiexec.c:2024): downstream was
killed by signal 11 (Segmentation fault)

Here is what the sbatch script looks like:


#SBATCH --job-name=mpihello-intelmpi
#SBATCH --output=mpihello-intelmpi.out.%j
#SBATCH --error=mpihello-intelmpi.err.%j

#SBATCH -N 3 --tasks-per-node=4
#SBATCH --time=10:00
#SBATCH --mem-per-cpu=100


To simplify things, I have tried launching a simple 'hostname' command in
place of mpihello-intelmpi, but I get the same segmentation fault issue.
Addtionally, I can directly launch the simple mpihello program as well as a
'hostname' command on the command line without issue.

Any ideas on what is going on?
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20190206/c08c0eb9/attachment.html>

More information about the slurm-users mailing list