[slurm-users] Stopping new jobs but letting old ones end
Ole Holm Nielsen
Ole.H.Nielsen at fysik.dtu.dk
Tue Feb 1 07:19:30 UTC 2022
One ting to be aware about when setting partition states to down:
* Setting partition state=down will be reset if slurmctld is restarted.
Read the slurmctld man-page under the -R parameter. So it's better not to
restart slurmctld during the downtime.
/Ole
On 2/1/22 08:11, Ole Holm Nielsen wrote:
> Login nodes being down doesn't affect Slurm jobs at all (except if you run
> slurmctld/slurmdbd on the login node ;-)
>
> To stop new jobs from being scheduled for running, mark all partitions
> down. This is useful when recovering the cluster from a power or cooling
> downtime, for example.
>
> I wrote a handy little script "schedjobs down" available from
> https://github.com/OleHolmNielsen/Slurm_tools/tree/master/jobs
> This loops over all partitions in the cluster and marks them down. When
> the cluster is OK again, run "schedjobs up".
>
> /Ole
>
> On 2/1/22 07:14, Sid Young wrote:
>> Brian / Christopher, that looks like a good process, thanks guys, I will
>> do some testing and let you know.
>>
>> if I mark a partition down and it has running jobs, what happens to
>> those jobs, do they keep running?
>>
>>
>> Sid Young
>> W: https://off-grid-engineering.com <https://off-grid-engineering.com>
>> W: (personal) https://sidyoung.com/ <https://sidyoung.com/>
>> W: (personal) https://z900collector.wordpress.com/
>> <https://z900collector.wordpress.com/>
>>
>>
>> On Tue, Feb 1, 2022 at 3:27 PM Brian Andrus <toomuchit at gmail.com
>> <mailto:toomuchit at gmail.com>> wrote:
>>
>> One possibility:
>>
>> Sounds like your concern is folks with interactive jobs from the login
>> node that are running under screen/tmux.
>>
>> That being the case, you need running jobs to end and not allow new
>> users to start tmux sessions.
>>
>> Definitely doing 'scontrol update state=down partition=xxxx' for each
>> partition. Also:
>>
>> touch /etc/nologin
>>
>> That will prevent new logins.
>>
>> Send a message to all active folks
>>
>> wall "system going down at XX:XX, please end your sessions"
>>
>> Then wait for folks to drain off your login node and do your stuff.
>>
>> When done, remove the /etc/nologin file and folks will be able to
>> login again.
>>
>> Brian Andrus
>>
>> On 1/31/2022 9:18 PM, Sid Young wrote:
>>>
>>>
>>>
>>> Sid Young
>>> W: https://off-grid-engineering.com <https://off-grid-engineering.com>
>>> W: (personal) https://sidyoung.com/ <https://sidyoung.com/>
>>> W: (personal) https://z900collector.wordpress.com/
>>> <https://z900collector.wordpress.com/>
>>>
>>>
>>> On Tue, Feb 1, 2022 at 3:02 PM Christopher Samuel <chris at csamuel.org
>>> <mailto:chris at csamuel.org>> wrote:
>>>
>>> On 1/31/22 4:41 pm, Sid Young wrote:
>>>
>>> > I need to replace a faulty DIMM chim in our login node so I
>>> need to stop
>>> > new jobs being kicked off while letting the old ones end.
>>> >
>>> > I thought I would just set all nodes to drain to stop new jobs
>>> from
>>> > being kicked off...
>>>
>>> That would basically be the way, but is there any reason why
>>> compute
>>> jobs shouldn't start whilst the login node is down?
>>>
>>>
>>> My concern was to keep the running jobs going and stop new jobs, so
>>> when the last running job ends,
>>> I could reboot the login node knowing that any terminal windows
>>> "screen"/"tmux" sessions would effectively
>>> have ended as the job(s) had now ended
>>>
>>> I'm not sure if there was an accepted procedure or best practice way
>>> to tackle shutting down the Login node for this use case.
>>>
>>> On the bright side I am down to two jobs left so any day now :)
More information about the slurm-users
mailing list