[slurm-users] Fairshare config change affect on running/queued jobs?
Walsh, Kevin
walsh at njit.edu
Fri Apr 30 16:39:37 UTC 2021
Hello everyone,
We wish to deploy "fair share" scheduling configuration and would like to
inquire if we should be aware of effects this might have on jobs already
running or already queued when the config is changed.
The proposed changes are from the example at
https://slurm.schedmd.com/archive/slurm-18.08.9/priority_multifactor.html#config
:
> # Activate the Multi-factor Job Priority Plugin with decay
> PriorityType=priority/multifactor
> # 2 week half-life
> PriorityDecayHalfLife=14-0
> # The larger the job, the greater its job size priority.
> PriorityFavorSmall=NO
> # The job's age factor reaches 1.0 after waiting in the
> # queue for 2 weeks.
> PriorityMaxAge=14-0
> # This next group determines the weighting of each of the
> # components of the Multi-factor Job Priority Plugin.
> # The default value for each of the following is 1.
> PriorityWeightAge=1000
> PriorityWeightFairshare=10000
> PriorityWeightJobSize=1000
> PriorityWeightPartition=1000
> PriorityWeightQOS=0 # don't use the qos factor
We're running SLURM 18.08.8 on CentOS Linux 7.8.2003. The current
slurm.conf is defaults as far as fair share is concerned:
> EnforcePartLimits=ALL
> GresTypes=gpu
> MpiDefault=pmix
> ProctrackType=proctrack/cgroup
> PrologFlags=x11,contain
> PropagateResourceLimitsExcept=MEMLOCK,STACK
> RebootProgram=/sbin/reboot
> ReturnToService=1
> SlurmctldPidFile=/var/run/slurmctld.pid
> SlurmctldPort=6817
> SlurmdPidFile=/var/run/slurmd.pid
> SlurmdPort=6818
> SlurmdSpoolDir=/var/spool/slurmd
> SlurmUser=slurm
> SlurmdSyslogDebug=verbose
> StateSaveLocation=/var/spool/slurm/ctld
> SwitchType=switch/none
> TaskPlugin=task/cgroup,task/affinity
> TaskPluginParam=Sched
> HealthCheckInterval=300
> HealthCheckProgram=/usr/sbin/nhc
> InactiveLimit=0
> KillWait=30
> MinJobAge=300
> SlurmctldTimeout=120
> SlurmdTimeout=300
> Waittime=0
> DefMemPerCPU=1024
> FastSchedule=1
> SchedulerType=sched/backfill
> SelectType=select/cons_res
> SelectTypeParameters=CR_Core_Memory
> AccountingStorageHost=sched-db.lan
> AccountingStorageLoc=slurm_acct_db
> AccountingStoragePass=/var/run/munge/munge.socket.2
> AccountingStoragePort=6819
> AccountingStorageType=accounting_storage/slurmdbd
> AccountingStorageUser=slurm
> AccountingStoreJobComment=YES
> AccountingStorageTRES=gres/gpu
> JobAcctGatherFrequency=30
> JobAcctGatherType=jobacct_gather/linux
> SlurmctldDebug=info
> SlurmdDebug=info
> SlurmSchedLogFile=/var/log/slurm/slurmsched.log
> SlurmSchedLogLevel=1
Node and partition configs are omitted above.
Any and all advice will be greatly appreciated.
Best wishes,
~Kevin
Kevin Walsh
Senior Systems Administration Specialist
New Jersey Institute of Technology
Academic & Research Computing Systems
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20210430/f6410150/attachment.htm>
More information about the slurm-users
mailing list