[slurm-users] Increasing job priority based on resources requested.
Chris M. Miller
cmmiller at cs.princeton.edu
Fri Apr 19 15:39:01 UTC 2019
Prentice,
I don't have a good answer to your original question, but I'll note I have a similar concern and solved it a different way. What we did was use lower weights in the node definitions for the "smaller" (less feature-rich) nodes, and extra high weights for nodes with unique features (like GPUs, in our case). In this way, jobs are scheduled to the smallest available nodes they fit in, and the larger or more feature-rich nodes have a kind of soft reservation either for large jobs or for busy times.
Cheers,
Chris
----- Original Message -----
From: "Prentice Bisbal" <pbisbal at pppl.gov>
To: slurm-users at lists.schedmd.com
Sent: Friday, April 19, 2019 11:27:08 AM
Subject: Re: [slurm-users] Increasing job priority based on resources requested.
Ryan,
I certainly understand your point of view, but yes, this is definitely
what I want. We only have a few large memory nodes, so we want jobs that
request a lot of memory to have higher priority so they get assigned to
those large memory nodes ahead of lower-memory jobs which could run
anywhere else. But we don't want those nodes to sit idle if there's jobs
in the queue that need that much memory. Similar idea for IB - nodes
that need IB should get priority over nodes that don't
Ideally, I wouldn't have such a heterogeneous environment, and then this
wouldn't be needed at all.
I agree this opens another avenue for unscrupulous users to game the
system, but that (in theory) can be policed by looked at memory
requested vs. memory used in the accounting data to identify any abusers
and then give them a stern talking to.
Prentice
On 4/18/19 5:27 PM, Ryan Novosielski wrote:
> This is not an official answer really, but I’ve always just considered this to be the way that the scheduler works. It wants to get work completed, so it will have a bias toward doing what is possible vs. not (can’t use 239GB of RAM on a 128GB node). And really, is a higher priority what you want? I’m not so sure. How soon will someone figure out that they might get a higher priority based on requesting some feature they don’t need?
>
> --
> ____
> || \\UTGERS, |---------------------------*O*---------------------------
> ||_// the State | Ryan Novosielski - novosirj at rutgers.edu
> || \\ University | Sr. Technologist - 973/972.0922 (2x0922) ~*~ RBHS Campus
> || \\ of NJ | Office of Advanced Research Computing - MSB C630, Newark
> `'
>
>> On Apr 18, 2019, at 5:20 PM, Prentice Bisbal <pbisbal at pppl.gov> wrote:
>>
>> Slurm-users,
>>
>> Is there away to increase a jobs priority based on the resources or constraints it has requested?
>>
>> For example, we have a very heterogeneous cluster here: Some nodes only have 1 Gb Ethernet, some have 10 Gb Ethernet, and others have DDR IB. In addition, we have some large memory nodes with RAM amounts ranging from 128 GB up to 512 GB. To allow a user to request IB, I have implemented that as a feature in the node definition so users can request that as a constraint.
>>
>> I would like to make it that if a job request IB, it's priority will go up, or if it requests a lot of memory (specifically memory-per-cpu), it's priority will go up proportionately to the amount of memory requested. Is this possible? If so, how?
>>
>> I have tried going through the documentation, and googling, but 'priority' is used to discuss job priority so much, I couldn't find any search results relevant to this.
>>
>> --
>> Prentice
>>
>>
More information about the slurm-users
mailing list