[slurm-users] possible to set memory slack space before killing jobs?

Bjørn-Helge Mevik b.h.mevik at usit.uio.no
Fri Dec 7 03:15:40 MST 2018


Eli V <eliventer at gmail.com> writes:

> On Wed, Dec 5, 2018 at 5:04 PM Bjørn-Helge Mevik <b.h.mevik at usit.uio.no> wrote:
>>
>> I don't think Slurm has any facility for soft memory limits.
>>
>> But you could emulate it by simply configure the nodes in slurm.conf
>> with, e.g., 15% higher RealMemory value than what is actually available
>> on the node.  Then a node with 256 GiB RAM would be able to run 9 jobs,
>> each asking for 32 GiB RAM.
>>
>> (You wouldn't get the effect that a job would be allowed to exceed its
>> soft limit for a set amount of time before getting killed, though.)
>
> I don't think this is possible currently. From my experience slurm
> will auto drain a node if it's actual physical memory is less then
> what's defined for it in the slurm.conf.

True.  I forgot about that.  You could prevent slurm from draining them
by setting FastSchedule=2 in slurm.conf, but then you wouldn't detect
nodes losing RAM (which does happen from time to time).

-- 
Regards,
Bjørn-Helge Mevik, dr. scient,
Department for Research Computing, University of Oslo
-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 832 bytes
Desc: not available
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20181207/19c5ac7b/attachment.sig>


More information about the slurm-users mailing list