[slurm-users] Submit Slurm Job in PE environment
Anson Abraham
anson.abraham at gmail.com
Thu Jun 21 10:09:35 MDT 2018
nevermind... after reading my email, i just realized the typo.
On Thu, Jun 21, 2018 at 11:22 AM Anson Abraham <anson.abraham at gmail.com>
wrote:
> Thanks,
> However, i'm on ubuntu, so when i installed w/ apt install slurm-llnl and
> slurm-llnl-basic-plugins{-dev}
>
> i added the con_res entry. however, the controller can't find the
> select/con_res plugin. Is there something else i need to install to get
> the plugin to be recognized?
>
> On Wed, Jun 20, 2018 at 2:49 AM Thomas HAMEL <hmlth at t-hamel.fr> wrote:
>
>> Hello,
>>
>> Slurm can have a behavior similar to the one you were used on SGE, it's
>> not the default but it's really commonly used. This is called Consumable
>> Resources in Slurm, you don't allocate full nodes but resources on the
>> node like cores or memory. It's activated in slurm.conf (after
>> restarting or reconfiguring the controller) :
>>
>> SelectType=select/cons_res
>> SelectTypeParameters=CR_Core_Memory
>>
>> https://slurm.schedmd.com/cons_res.html
>>
>> MPI serves another purpose that do not really fit you model. The
>> solution to your issue is Consumable Resources but you can take a look
>> at Job Arrays to group all jobs of your loop together.
>>
>> Regards,
>>
>> Thomas HAMEL
>>
>>
>> On 19/06/2018 23:33, Anson Abraham wrote:
>> > Hi,
>> > Relatively new to Slurm. I've been using Sun GridEngine mostly.
>> > I have a cluster of 3 machines each with 8 cores. In SGE i allocate
>> > the PE slots per machine, where if i submit 24 jobs it'll run all 24
>> > jobs (b/c each job will use 1 core).
>> > however, if i submit job in Slurm through sbatch i can only get it to
>> > run 3 jobs at a time, even when i define the cpus_per_task. I was
>> > told to use openMPI for this.
>> > i'm not familiar w/ openMPI so i did an apt install of libopenmpi-dev.
>> >
>> > Do i have to loop through my job submission w/ mpirun
>> > and run an sbatch outside of it?
>> > Again, i'm still new to this, and w/ sge it was pretty straight fwd
>> > where all i had to do was:
>> > loop through files
>> > qsub -N {name of job} script.sh {filename}.
>> >
>> > not sure how would i do so here.
>>
>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20180621/f770bda9/attachment.html>
More information about the slurm-users
mailing list