[slurm-users] [External] Submitting to multiple paritions problem with gres specified
ewan.roche at unil.ch
Tue Mar 9 08:37:16 UTC 2021
as a variant on what has already been suggested we also have the CPU type as a feature:
This allows people that want the same architecture and not just the same instruction set for a multi-node job can say:
Apart from multiple partitions approach another hack/workaround is to abuse the topology plugin to create fake switches with nodes of each CPU type connected and no links between these switches.
As there is no link between these “switches” Slurm will never schedule a job on node01 and node03.
Division Calcul et Soutien à la Recherche
UNIL | Université de Lausanne
> On 9 Mar 2021, at 09:16, Ward Poelmans <ward.poelmans at vub.be> wrote:
> Hi Prentice,
> On 8/03/2021 22:02, Prentice Bisbal wrote:
>> I have a very hetergeneous cluster with several different generations of
>> AMD and Intel processors, we use this method quite effectively.
> Could you elaborate a bit more and how you manage that? Do you force you
> users to pick a feature? What if a user submits a multi node job, can
> you make sure it will not start over a mix of avx512 and avx2 nodes?
>> If you want to continue down the road you've already started on, can you
>> provide more information, like the partition definitions and the gres
>> definitions? In general, Slurm should support submitting to multiple
> As far as I understood it, you can give a comma separated list of
> partitions to sbatch but it's not possible to this by default?
More information about the slurm-users