[slurm-users] srun --mem issue

Felho, Sandor Sandor.Felho at transunion.com
Thu Dec 8 14:51:29 UTC 2022


Bill - thank you for the code.

For the record, our queues explicitly block any requests greater than what the queue allows - MaxMemPerNode = 64 GiB's.

If there are any other ideas, feel free to comment.

Sandor
________________________________
From: slurm-users <slurm-users-bounces at lists.schedmd.com> on behalf of Bill <bill at princeton.edu>
Sent: Thursday, December 8, 2022 7:52 AM
To: slurm-users at lists.schedmd.com <slurm-users at lists.schedmd.com>
Subject: Re: [slurm-users] srun --mem issue

CAUTION:   This email originated from outside of the organization. Do not click links or open attachments unless you can confirm the sender and know the content is safe.

If you use a job_submit.lua script just add

if job_desc.pn_min_memory == 0 or job_desc.min_mem_per_cpu == 0 then
            log_info("slurm_job_submit: job from uid %d invalid memory
request MaxMemPerNode", job_desc.user_id)
            return 2044 -- signal ESLURM_INVALID_TASK_MEMORY
end


Bill

On 12/7/22 12:03 PM, Felho, Sandor wrote:
> TransUnion is running a ten-node site using slurm with multiple queues.
> We have an issue with --mem parameter. The is one user who has read the
> slurm manual and found the --mem=0. This is giving the maximum memory on
> the node (500 GiB's) for the single job. How can I block a --mem=0 request?
>
> We are running:
>
>   * OS: RHEL 7
>   * cgroups version 1
>   * slurm: 19.05
>
> Thank you,
>
> *Sandor Felho *
>
> Sr Consultant, Data Science & Analytics
>
>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20221208/a0fbe98e/attachment.htm>


More information about the slurm-users mailing list