[slurm-users] RawUsage 0??
matthias.leopold at meduniwien.ac.at
Wed Apr 7 10:52:24 UTC 2021
The "solution" for my problem was very simple: after reboot of all hosts
in this test cluster (login node with slurmctld/slurmdbd + 2 worker
nodes) I do get reasonable values in sshare. Maybe I didn't do that
after finishing deepops installation procedure (but I didn't know I had
to do it and had no hints)
Sorry for bothering you
Am 06.04.21 um 17:06 schrieb Matthias Leopold:
> I'm very new to Slurm and try to understand basic concepts. One of them
> is the "Multifactor Priority Plugin". For this I submitted some jobs and
> looked at sshare output. To my surprise I don't get any numbers for
> "RawUsage", regardless what I do RawUsage stays 0 (same in "scontrol
> show assoc_mgr" output). When I look at CPU stats for the jobs I submit
> and complete (with sacct) I do see usage counters there, I also see
> counters for TRESRunMins in sshare while the job is running, but
> RawUsage (and also GrpTRESRaw) stays empty.
> I found this discussion for a similar topic:
> I can confirm that I waited for longer than 5 min for sshare to update
> values and I also tried "PriorityDecayHalfLife=0".
> My slurm.conf currently includes:
> JobAcctGatherType=jobacct_gather/linux # also tried "cgroup"
> # only partition
> PartitionName=batch Nodes=ALL Default=YES DefMemPerCPU=0 State=UP
> OverSubscribe=NO MaxTime=INFINITE
> Slurm is 20.11.3, installed via NVIDIA/deepops
> (https://github.com/nvidia/deepops/) ansible playbooks in VMs.
> What am I missing?
> Thanks for any advice
More information about the slurm-users