[slurm-users] SLURM 17.02.9 slurmctld unresponsive with server_thread_count over limit, waiting in syslog
Paul Edmon
pedmon at cfa.harvard.edu
Wed Nov 8 13:57:49 MST 2017
So hangups like this can occur due to the slurmdbd being busy with
requests. I've seen that happen when an ill timed massive sacct request
hits when slurmdbd is doing its roll up. In that case the slurmctld
hangs while slurmdbd is busy. Typically in this case restarting
mysql/slurmdbd seems to fix the issue.
Otherwise this can happen due to massive traffic to the slurmctld. You
can try using the defer option for the SchedulerParamters. That slows
down the scheduler so it can handle the additional load.
-Paul Edmon-
On 11/8/2017 3:11 PM, Sean Caron wrote:
> Hi all,
>
> I see SLURM 17.02.9 slurmctld hang or become unresponsive every few
> days with the message in syslog:
>
> server_thread_count over limit (256), waiting
>
> I believe from the user perspective they see "Socket timed out on
> send/recv operation". Slurmctld never seems to recover once it's in
> this state and will not respond to /etc/init.d/slurm restart. Only
> after an admin does a kill -9 and restarts slurmctld does it snap back.
>
> I don't see anything else in the logs that looks like an error message
> that would help diagnose what is going on, even with log level debug3
> on the SLURM controller daemon.
>
> I monitor CPU and memory utilization with "htop" on the machine
> running the controller daemon and it doesn't seem like it's
> overwhelmed by slurmctld load or anything like that.
>
> Machine running the controller daemon feels reasonable for the task,
> for the size of our cluster. It's a repurposed Dell PowerEdge R410
> with 24 threads and 32 GB physical. Unless I'm way off?
>
> I tried all kinds of SchedulerParameter tweaks on sched/backfill and
> even set the scheduler back to sched/builtin and it's still happening.
> Didn't seem to affect the frequency much, either.
>
> Any thoughts what could be causing SLURM to spawn so many threads and
> hang up?
>
> Our cluster is medium-sized, we probably have a few thousand jobs in
> the queue on average at any given time.
>
> Monitoring with sdiag, the max cycle time of main scheduler never
> cracks 2 seconds. This seems reasonable?
>
> Thanks,
>
> Sean
>
More information about the slurm-users
mailing list