[slurm-users] SLURM heterogeneous jobs, a little help needed plz
    Frava 
    fravadona at gmail.com
       
    Tue Mar 19 21:03:27 UTC 2019
    
    
  
Hi all,
I'm struggling to get an heterogeneous job to run...
The SLURM version installed on the cluster is 17.11.12
Here are the SBATCH file parameters of the job :
================================================================================
#!/bin/bash
#SBATCH --threads-per-core=1
#SBATCH --ntasks=1
#SBATCH --cpus-per-task=1
#SBATCH --mem-per-cpu=8G
#
#SBATCH packjob
#
#SBATCH --threads-per-core=1
#SBATCH --cpus-per-task=4
#SBATCH --mem-per-cpu=2G
#SBATCH --nodes=2
#SBATCH --ntasks-per-node=2
#SBATCH --gres=gpu:2
#SBATCH --gres-flags=enforce-binding
================================================================================
Now the SRUN commands that I tried in the SBATCH file:
1) srun --mpi=pmix myapp
=> The app gets allocated only 1 MPI rank
2) srun --mpi=pmix --pack-group=0 --ntasks=1 : --pack-group=1 --ntasks=4
myapp
=> srun: fatal: Job steps that span multiple components of a heterogeneous
job are not currently supported
3) srun --mpi=pmix --pack-group=0 --ntasks=1 myapp : --pack-group=1
--ntasks=4 myapp
=> srun: fatal: Job steps that span multiple components of a heterogeneous
job are not currently supported
So my question is: How do we run an heterogeneous job ?
Thanks for your tips,
Rafael N.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20190319/b64587fd/attachment-0001.html>
    
    
More information about the slurm-users
mailing list