[slurm-users] GrpTRESMins and GrpTRESRaw usage

gerard.gil at cines.fr gerard.gil at cines.fr
Tue Jun 28 17:30:58 UTC 2022


Hi Miguel, 

OK, I did'nt know this command. 

I'm not sure to understand how it works regarding to my goal. 
I use the following command inspired by the command you gave me and I obtain a UsageRaw for each QOS. 

scontrol -o show assoc_mgr -accounts=myaccount Users=" " 

Do I have to sumup all QOS RawUsage to obtain the RawUsage of myaccount with NoDecay ? 
If I set GrpTRESMins for an Account and not for a QOS, does SLURM handle to sumpup these QOS RawUsage to control if the GrpTRESMins account limit is reach ? 

Thanks again for your precious help. 

Gérard 
[ http://www.cines.fr/ ] 

> De: "Miguel Oliveira" <miguel.oliveira at uc.pt>
> À: "Slurm-users" <slurm-users at lists.schedmd.com>
> Envoyé: Mardi 28 Juin 2022 17:23:18
> Objet: Re: [slurm-users] GrpTRESMins and GrpTRESRaw usage

> Hi Gérard,

> The way you are checking is against the association and as such it ought to be
> decreasing in order to be used by fair share appropriately.
> The counter used that does not decrease is on the QoS, not the association. You
> can check that with:

> scontrol -o show assoc_mgr | grep "^QOS='+account+’ ”

> That ought to give you two numbers. The first is the limit, or N for not limit,
> and the second in parenthesis the usage.

> Hope that helps.

> Best,

> Miguel Afonso Oliveira

>> On 28 Jun 2022, at 08:58, [ mailto:gerard.gil at cines.fr | gerard.gil at cines.fr ]
>> wrote:

>> Hi Miguel,

>> I modified my test configuration to evaluate the effect of NoDecay.

>> I modified all QOS adding NoDecay Flag.

>> toto at login1:~/TEST$ sacctmgr show QOS
>> Name Priority GraceTime Preempt PreemptExemptTime PreemptMode Flags UsageThres
>> UsageFactor GrpTRES GrpTRESMins GrpTRESRunMin GrpJobs GrpSubmit GrpWall MaxTRES
>> MaxTRESPerNode MaxTRESMins MaxWall MaxTRESPU MaxJobsPU MaxSubmitPU MaxTRESPA
>> MaxJobsPA MaxSubmitPA MinTRES
>> ---------- ---------- ---------- ---------- ------------------- -----------
>> ---------------------------------------- ---------- ----------- -------------
>> ------------- ------------- ------- --------- ----------- -------------
>> -------------- ------------- ----------- ------------- --------- -----------
>> ------------- --------- ----------- -------------
>> normal 0 00:00:00 cluster NoDecay 1.000000
>> interactif 10 00:00:00 cluster NoDecay 1.000000 node=50 node=22 1-00:00:00
>> node=50
>> petit 4 00:00:00 cluster NoDecay 1.000000 node=1500 node=22 1-00:00:00 node=300
>> gros 6 00:00:00 cluster NoDecay 1.000000 node=2106 node=700 1-00:00:00 node=700
>> court 8 00:00:00 cluster NoDecay 1.000000 node=1100 node=100 02:00:00 node=300
>> long 4 00:00:00 cluster NoDecay 1.000000 node=500 node=200 5-00:00:00 node=200
>> special 10 00:00:00 cluster NoDecay 1.000000 node=2106 node=2106 5-00:00:00
>> node=2106
>> support 10 00:00:00 cluster NoDecay 1.000000 node=2106 node=700 1-00:00:00
>> node=2106
>> visu 10 00:00:00 cluster NoDecay 1.000000 node=4 node=700 06:00:00 node=4

>> I submitted a bunch of jobs to control the NoDecay efficiency and I noticed
>> RawUsage as well as GrpTRESRaw cpu is still decreasing.

>> toto at login1:~/TEST$ sshare -A dci -u " " -o account,user,GrpTRESRaw%80,
>> GrpTRESMins ,RawUsage
>> Account User GrpTRESRaw GrpTRESMins RawUsage
>> -------------------- ----------
>> -----------------------------------------------------
>> ------------------------------ -----------
>> dci cpu=6932
>> ,mem=12998963,energy=0,node=216,billing=6932,fs/disk=0,vmem=0,pages=0 cpu=17150
>> 415966
>> toto at login1:~/TEST$ sshare -A dci -u " " -o account,user,GrpTRESRaw%80,
>> GrpTRESMins , RawUsage
>> Account User GrpTRESRaw GrpTRESMins RawUsage
>> -------------------- ----------
>> -----------------------------------------------------
>> ------------------------------ -----------
>> dci cpu=6931
>> ,mem=12995835,energy=0,node=216,billing=6931,fs/disk=0,vmem=0,pages=0 cpu=17150
>> 415866
>> toto at login1:~/TEST$ sshare -A dci -u " " -o
>> account,user,GrpTRESRaw%80,GrpTRESMins,RawUsage
>> Account User GrpTRESRaw GrpTRESMins RawUsage
>> -------------------- ----------
>> -----------------------------------------------------
>> ------------------------------ -----------
>> dci cpu=6929
>> ,mem=12992708,energy=0,node=216,billing=6929,fs/disk=0,vmem=0,pages=0 cpu=17150
>> 415766

>> Something I forgot to do ?

>> Best,
>> Gérard

>> Cordialement,
>> Gérard Gil

>> Département Calcul Intensif
>> Centre Informatique National de l'Enseignement Superieur
>> 950, rue de Saint Priest
>> 34097 Montpellier CEDEX 5
>> FRANCE

>> tel : (334) 67 14 14 14
>> fax : (334) 67 52 37 63
>> web : [ http://www.cines.fr/ | http://www.cines.fr ]

>>> De: "Gérard Gil" < [ mailto:gerard.gil at cines.fr | gerard.gil at cines.fr ] >
>>> À: "Slurm-users" < [ mailto:slurm-users at lists.schedmd.com |
>>> slurm-users at lists.schedmd.com ] >
>>> Cc: "slurm-users" < [ mailto:slurm-users at schedmd.com | slurm-users at schedmd.com ]
>>> >
>>> Envoyé: Vendredi 24 Juin 2022 14:52:12
>>> Objet: Re: [slurm-users] GrpTRESMins and GrpTRESRaw usage

>>> Hi Miguel,

>>> Good !!

>>> I'll try this options on all existing QOS and see if everything works as
>>> expected.
>>> I'll inform you on the results.

>>> Thanks a lot

>>> Best,
>>> Gérard

>>> ----- Mail original -----

>>>> De: "Miguel Oliveira" < [ mailto:miguel.oliveira at uc.pt | miguel.oliveira at uc.pt ]
>>>> >
>>>> À: "Slurm-users" < [ mailto:slurm-users at lists.schedmd.com |
>>>> slurm-users at lists.schedmd.com ] >
>>>> Cc: "slurm-users" < [ mailto:slurm-users at schedmd.com | slurm-users at schedmd.com ]
>>>> >
>>>> Envoyé: Vendredi 24 Juin 2022 14:07:16
>>>> Objet: Re: [slurm-users] GrpTRESMins and GrpTRESRaw usage
>>>> Hi Gérard,

>>>> I believe so. All our accounts correspond to one project and all have an
>>>> associated QoS with NoDecay and DenyOnLimit. This is enough to restrict usage
>>>> on each individual project.
>>>> You only need these flags on the QoS. The association will carry on as usual and
>>>> fairshare will not be impacted.

>>>> Hope that helps,

>>>> Miguel Oliveira

>>>>> On 24 Jun 2022, at 12:56, [ mailto:gerard.gil at cines.fr | gerard.gil at cines.fr ]
>>>>> wrote:

>>>>> Hi Miguel,

>>>>>> Why not? You can have multiple QoSs and you have other techniques to change
>>>>>> priorities according to your policies.
>>>>> Is this answer my question ?

>>>>> "If all configured QOS use NoDecay, we can take advantage of the FairShare
>>>>> priority with Decay and all jobs GrpTRESRaw with NoDecay ?"

>>>>> Thanks

>>>>> Best,
>>> > > Gérard
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20220628/cde92c90/attachment-0001.htm>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: smime.p7s
Type: application/pkcs7-signature
Size: 2065 bytes
Desc: S/MIME Cryptographic Signature
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20220628/cde92c90/attachment-0001.bin>


More information about the slurm-users mailing list