[slurm-users] Defining constraints for job dispatching

Mahmood Naderan mahmood.nt at gmail.com
Thu Sep 20 01:50:54 MDT 2018


Hi Michael,
Sorry for the late response. Do you mean supplying --exclusive to the
srun command? Or I have to do something else for partitions? Currently
they use

srun -n 1 -c 6 --x11 -A monthly -p CAT --mem=32GB ./fluent.sh

where fluent.sh is

#!/bin/bash
unset SLURM_GTIDS
/state/partition1/ansys_inc/v140/fluent/bin/fluent


Regards,
Mahmood




On Sat, Sep 1, 2018 at 7:45 PM Renfro, Michael <Renfro at tntech.edu> wrote:
>
> Depending on the scale (what percent are Fluent users, how many nodes you have), you could use exclusive mode on either a per-partition or per-job basis.
>
> Here, my (currently few) Fluent users do all their GUI work off the cluster, and just submit batch jobs using the generated case and data files.
>
> --
> Mike Renfro  / HPC Systems Administrator, Information Technology Services
> 931 372-3601 / Tennessee Tech University
>
> > On Sep 1, 2018, at 9:53 AM, Mahmood Naderan <mahmood.nt at gmail.com> wrote:
> >
> > Hi,
> > I have found that when user A is running a fluent job (some 100% processes in top) and user B decides to run a fluent job for his own, the console window of fluent shows some messages that another fluent process is running and it can not set affinity. This is not an error, but I see that the speed is somehow low.
> >
> > Think that when a user runs "srun --x11 .... script" where script launches some fluent processes and slurm put that job on compute-0-0, there should be a way that another "script" from another user goes to compute-0-1 even if compute-0-0 has free cores.
> >
> > Is there any way in slurm configuration to set such a constraint? If slurm wants to dispatch a job, first see if process X is running there or not.
> >
> >
> > Regards,
> > Mahmood
> >
> >
>
>



More information about the slurm-users mailing list