[slurm-users] Advice on using GrpTRESRunMin=cpu=<limit>
    David Baker 
    D.J.Baker at soton.ac.uk
       
    Wed Feb 12 16:45:32 UTC 2020
    
    
  
Hello,
Before implementing "GrpTRESRunMin=cpu=limit" on our production cluster I'm doing some tests on the development cluster. I've only get a handful of compute nodes to play without and so I have set the limit sensibly low. That is, I've set the limit to be 576,000. That's equivalent to 400 CPU-days. In other words, I can potentially submit the following job...
1 x 2 nodes x 80 cpus/node x 2.5 days = 400 CPU-days
I submitted a set of jobs requesting 2 nodes, 80 cpus/node for 2.5 days. The first day is running and the rest are in the queue -- what I see makes sense...
JOBID PARTITION     NAME     USER ST       TIME  NODES NODELIST(REASON)
677     debug    myjob     djb1 PD       0:00      2 (AssocGrpCPURunMinutesLimit)
678     debug    myjob     djb1 PD       0:00      2 (AssocGrpCPURunMinutesLimit)
679     debug    myjob     djb1 PD       0:00      2 (AssocGrpCPURunMinutesLimit)
676     debug    myjob     djb1  R      12:52      2 navy[54-55]
On the other hand, I expected these jobs not to accrue priority, however they do appear to be (see sprio below). I'm working with Slurm v19.05.2. Have I missed something vital/important in the config? We hoped that the queued jobs would not accrue priority. We haven't, for example, used "accrue always". Have I got that wrong? Could someone please advise us.
Best regards,
David
[root at navy51 slurm]# sprio
          JOBID PARTITION   PRIORITY       SITE        AGE  FAIRSHARE    JOBSIZE        QOS
            677 debug        5551643     100000       1644     450000    5000000          0
            678 debug        5551643     100000       1644     450000    5000000          0
            679 debug        5551642     100000       1643     450000    5000000          0
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20200212/bb2a82a0/attachment-0001.htm>
    
    
More information about the slurm-users
mailing list