[slurm-users] Slurm 17.11 and configuring backfill and oversubscribe to allow concurrent processes

Renfro, Michael Renfro at tntech.edu
Thu Feb 27 17:44:19 UTC 2020


If that 32 GB is main system RAM, and not GPU RAM, then yes. Since our GPU nodes are over-provisioned in terms of both RAM and CPU, we end up using the excess resources for non-GPU jobs.

If that 32 GB is GPU RAM, then I have no experience with that, but I suspect MPS would be required.

> On Feb 27, 2020, at 11:14 AM, Robert Kudyba <rkudyba at fordham.edu> wrote:
> 
> So looking at the new cons_tres option at https://slurm.schedmd.com/SLUG19/GPU_Scheduling_and_Cons_Tres.pdf, would we be able to use, e.g., --mem-per-gpu= Memory per allocated GPU, and it a user allocated --mem-per-gpu=8, and the V100 we have is 32 GB, will subsequent jobs be able to use the remaining 24 GB?




More information about the slurm-users mailing list