[slurm-users] Submit Slurm Job in PE environment

Anson Abraham anson.abraham at gmail.com
Thu Jun 21 09:22:34 MDT 2018

However, i'm on ubuntu, so when i installed w/ apt install slurm-llnl and

i added the con_res entry.  however, the controller can't find the
select/con_res plugin.  Is there something else i need to install to get
the plugin to be recognized?

On Wed, Jun 20, 2018 at 2:49 AM Thomas HAMEL <hmlth at t-hamel.fr> wrote:

> Hello,
> Slurm can have a behavior similar to the one you were used on SGE, it's
> not the default but it's really commonly used. This is called Consumable
> Resources in Slurm, you don't allocate full nodes but resources on the
> node like cores or memory. It's activated in slurm.conf (after
> restarting or reconfiguring the controller) :
> SelectType=select/cons_res
> SelectTypeParameters=CR_Core_Memory
> https://slurm.schedmd.com/cons_res.html
> MPI serves another purpose that do not really fit you model. The
> solution to your issue is Consumable Resources but you can take a look
> at Job Arrays to group all jobs of your loop together.
> Regards,
> Thomas HAMEL
> On 19/06/2018 23:33, Anson Abraham wrote:
> > Hi,
> > Relatively new to Slurm.  I've been using Sun GridEngine mostly.
> > I have a cluster of 3 machines each with 8 cores.  In SGE i allocate
> > the PE slots per machine, where if i submit 24 jobs it'll run all 24
> > jobs (b/c each job will use 1 core).
> > however, if i submit job in Slurm through sbatch i can only get it to
> > run 3 jobs at a time, even when i define the cpus_per_task.  I was
> > told to use openMPI for this.
> > i'm not familiar w/ openMPI so i did an apt install of libopenmpi-dev.
> >
> > Do i have to loop through my job submission w/ mpirun
> > and run an sbatch outside of it?
> > Again, i'm still new to this, and w/ sge it was pretty straight fwd
> > where all i had to do was:
> >  loop through files
> >     qsub -N {name of job}  script.sh {filename}.
> >
> > not sure how would i do so here.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20180621/04131c19/attachment.html>

More information about the slurm-users mailing list