[slurm-users] Slurm 17.11 and configuring backfill and oversubscribe to allow concurrent processes
mercan
ahmet.mercan at uhem.itu.edu.tr
Thu Feb 27 08:53:01 UTC 2020
Hi;
At your partition definition, there is "Shared=NO". This is means "do
not share nodes between jobs". This parameter conflict with
"OverSubscribe=FORCE:12 " parameter. Acording to the slurm
documentation, the Shared parameter has been replaced by the
OverSubscribe parameter. But, I suppose it still works.
Regards,
Ahmet M.
On 26.02.2020 22:56, Robert Kudyba wrote:
> We run Bright 8.1 and Slurm 17.11. We are trying to allow for multiple
> concurrent jobs to run on our small 4 node cluster.
>
> Based on
> https://community.brightcomputing.com/question/5d6614ba08e8e81e885f1991?action=artikel&cat=14&id=410&artlang=en&highlight=slurm+%2526%252334%253Bgang+scheduling%2526%252334%253B
> and
> https://slurm.schedmd.com/cons_res_share.html
>
> Here are some settings in /etc/slurm/slurm.conf:
>
> SchedulerType=sched/backfill
> # Nodes
> NodeName=node[001-003] CoresPerSocket=12 RealMemory=191800 Sockets=2
> Gres=gpu:1
> # Partitions
> PartitionName=defq Default=YES MinNodes=1 AllowGroups=ALL
> PriorityJobFactor=1 PriorityTier=1 DisableRootJobs=NO RootOnly=NO
> Hidden=NO Shared=NO GraceTime=0 PreemptMode=OFF ReqResv=NO
> AllowAccounts=ALL AllowQos=ALL LLN=NO ExclusiveUser=NO
> OverSubscribe=FORCE:12 OverTimeLimit=0 State=UP Nodes=node[001-003]
> PartitionName=gpuq Default=NO MinNodes=1 AllowGroups=ALL
> PriorityJobFactor=1 PriorityTier=1 DisableRootJobs=NO RootOnly=NO
> Hidden=NO Shared=NO GraceTime= 0 PreemptMode=OFF ReqResv=NO
> AllowAccounts=ALL AllowQos=ALL LLN=NO ExclusiveUser=NO
> OverSubscribe=FORCE:12 OverTimeLimit=0 State=UP
> # Generic resources types
> GresTypes=gpu,mic
> # Epilog/Prolog parameters
> PrologSlurmctld=/cm/local/apps/cmd/scripts/prolog-prejob
> Prolog=/cm/local/apps/cmd/scripts/prolog
> Epilog=/cm/local/apps/cmd/scripts/epilog
> # Fast Schedule option
> FastSchedule=1
> # Power Saving
> SuspendTime=-1 # this disables power saving
> SuspendTimeout=30
> ResumeTimeout=60
> SuspendProgram=/cm/local/apps/cluster-tools/wlm/scripts/slurmpoweroff
> ResumeProgram=/cm/local/apps/cluster-tools/wlm/scripts/slurmpoweron
> # END AUTOGENERATED SECTION -- DO NOT REMOVE
> #
> http://kb.brightcomputing.com/faq/index.php?action=artikel&cat=14&id=410&artlang=en&highlight=slurm+%26%2334%3Bgang+scheduling%26%2334%3B
> SelectType=select/cons_res
> SelectTypeParameters=CR_CPU
> SchedulerTimeSlice=60
> EnforcePartLimits=YES
>
> But it appears each job takes 1 of the 3 nodes and all other jobs are
> back scheduled. Do we have an incorrect option set?
>
> squeue -a
> JOBID PARTITION NAME USER ST TIME NODES NODELIST(REASON)
> 1937 defq PaNet5 user1 PD 0:00 1 (Resources)
> 1938 defq PoNet5 user1 PD 0:00 1 (Priority)
> 1964 defq SENet5 user1 PD 0:00 1 (Priority)
> 1979 defq IcNet5 user1 PD 0:00 1 (Priority)
> 1980 defq runtrain user2 PD 0:00 1 (Priority)
> 1981 defq InRes5 user1 PD 0:00 1 (Priority)
> 1983 defq run_LSTM user3 PD 0:00 1 (Priority)
> 1984 defq run_hui. user4 PD 0:00 1 (Priority)
> 1936 defq SeRes5 user1 R 10:02:39 1 node003
> 1950 defq sequenti user5 R 1-02:03:00 1 node001
> 1978 defq run_hui. user16 R 13:48:21 1 node002
>
> Am I misunderstanding some of the settings?
>
>
More information about the slurm-users
mailing list