[slurm-users] Queue size, slow/unresponsive head node
Nicholas C Santucci
santucci at uci.edu
Thu Jan 11 22:25:22 MST 2018
Why do you have?
SchedulerParameters = (null)
Is that even allowed
?
https://slurm.schedmd.com/sched_config.html
On Thu, Jan 11, 2018 at 1:39 PM, Colas Rivière <riviere at umdgrb.umd.edu>
wrote:
> Hello,
>
> I'm managing a small cluster (one head node, 24 workers, 1160 total worker
> threads). The head node has two E5-2680 v3 CPUs (hyper-threaded), ~100 GB
> of memory and spinning disks.
> The head node becomes occasionally less responsive when there are more
> than 10k jobs in queue, and becomes really unmanageable when reaching 100k
> jobs in queue, with error messages such as:
>
>> sbatch: error: Slurm temporarily unable to accept job, sleeping and
>> retrying.
>>
> or
>
>> Running: slurm_load_jobs error: Socket timed out on send/recv operation
>>
> Is that normal to experience slowdowns when the queue reaches this few 10k
> jobs? What limit should I expect? Would adding a SSD drive for
> SlurmdSpoolDir help? What can be done to push this limit?
>
> The cluster runs Slurm 17.02.4 on CentOS 6 and the config is attached
> (from `scontrol show config`).
>
> Thanks,
> Colas
>
--
Nick Santucci
santucci at uci.edu
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20180111/2c75ec17/attachment-0003.html>
More information about the slurm-users
mailing list