[slurm-users] Limiting srun to a specific partition
Ewan Roche
ewan.roche at unil.ch
Tue Feb 15 10:13:01 UTC 2022
It doesn’t affect the use case of connecting via srun afterwards as no new job is submitted so the job_submit.lua logic is never called.
$ srun --pty /bin/bash
srun: error: submit_job: ERROR: interactive jobs are not allowed in the CPU or GPU partitions. Use the interactive partition
srun: error: Unable to allocate resources: Unspecified error
$ sbatch sleep.run
Submitted batch job 4624710
$ srun --jobid=4624710 --pty /bin/bash
[user at dna025 jobs]$ $ echo $TMPDIR
/tmp/4624710
Thanks
Ewan Roche
Division Calcul et Soutien à la Recherche
UNIL | Université de Lausanne
> On 15 Feb 2022, at 10:27, Tina Friedrich <tina.friedrich at it.ox.ac.uk> wrote:
>
> ...that would interfere with users 'logging in' to a job to check on it though, wouldn't it? I mean we do have pam_slurm_adopt configured but I still tell people it's preferable to use 'srun --jobid=XXXX --pty /bin/bash' to check what a specific job is doing as pam_slurm_adopt doesn't seem to inherit things like the job-local tmp dir and such :(.
>
> So I'd at least check if it's an existing JobID you're trying to connect to.
>
> Tina
>
> On 15/02/2022 08:08, Ewan Roche wrote:
>> Hi Peter,
>> as Rémi said, the way to do this in Slurm is via a job submit plugin. For example in our job_submit.lua we have
>> if (job_desc.partition == "cpu" or job_desc.partition == "gpu") and job_desc.qos ~= "admin" then
>> if job_desc.script == nil or job_desc.script == '' then
>> slurm.log_info("slurm_job_submit: jobscript is missing, assuming interactive job")
>> slurm.log_info("slurm_job_submit: CPU/GPU partition for interactive job, abort")
>> slurm.log_user("submit_job: ERROR: interactive jobs are not allowed in the CPU or GPU partitions. Use the interactive partition")
>> return -1
>> end
>> end
>> Which checks to see if the job script exists - this is more or less the definition of an interactive job.
>> Ewan Roche
>> Division Calcul et Soutien à la Recherche
>> UNIL | Université de Lausanne
>>> On 15 Feb 2022, at 08:47, Rémi Palancher <remi at rackslab.io> wrote:
>>>
>>> Hi Peter,
>>>
>>> Le lundi 14 février 2022 à 18:37, Peter Schmidt <pschmidt at rivosinc.com> a écrit :
>>>
>>>> slurm newbie here, converting from pbspro. In pbspro there is the capability of limiting interactive jobs (i.e srun) to a specific queue (i.e partition).
>>>
>>> Note that in Slurm, srun and interactive jobs are not the same things. The srun command is for creating steps of jobs (interactive or not), optionally creating a job allocation beforehand if it does not exist.
>>>
>>> You can run interactive jobs with salloc and even attach your PTY to a running batch job to interact with it. On the other hand, batchs jobs can create steps using srun command.
>>>
>>> I don't know any native Slurm feature to restrict interactive jobs (to a specific partition or whatever). However, using job_submit LUA plugin and a custom LUA script, you might be able to accomplish what you are expecting. It has been discussed here:
>>>
>>> https://bugs.schedmd.com/show_bug.cgi?id=3094
>>>
>>> Best,
>>> --
>>> Rémi Palancher
>>> Rackslab: Open Source Solutions for HPC Operations
>>> https://rackslab.io
>>>
>>>
>
> --
> Tina Friedrich, Advanced Research Computing Snr HPC Systems Administrator
>
> Research Computing and Support Services
> IT Services, University of Oxford
> http://www.arc.ox.ac.uk http://www.it.ox.ac.uk
>
More information about the slurm-users
mailing list