Hi Loris,
we use SLURM 23.02.7 (Production) and 23.11.1 (Testsystem). Our config contains a second parameter InteractiveStepOptions in slurm.conf:
InteractiveStepOptions="--interactive --preserve-env --pty $SHELL -l" LaunchParameters=enable_nss_slurm,use_interactive_step
That works fine for us:
[k202068@levantetest ~]$ salloc -N1 -A k20200 -p compute salloc: Pending job allocation 857 salloc: job 857 queued and waiting for resources salloc: job 857 has been allocated resources salloc: Granted job allocation 857 salloc: Waiting for resource configuration salloc: Nodes lt10000 are ready for job [k202068@lt10000 ~]$
Best Regards, Carsten
Am 05.09.24 um 14:17 schrieb Loris Bennett via slurm-users:
Hi,
With
$ salloc --version slurm 23.11.10
and
$ grep LaunchParameters /etc/slurm/slurm.conf LaunchParameters=use_interactive_step
the following
$ salloc --partition=interactive --ntasks=1 --time=00:03:00 --mem=1000 --qos=standard salloc: Granted job allocation 18928869 salloc: Nodes c001 are ready for job
creates a job
$ squeue --me JOBID PARTITION NAME USER ST TIME NODES NODELIST(REASON) 18928779 interacti interact loris R 1:05 1 c001
but causes the terminal to block.
From a second terminal I can log into the compute node:
$ ssh c001 [13:39:36] loris@c001 (1000) ~
Is that the expected behaviour or should salloc return a shell directly on the compute node (like srun --pty /bin/bash -l used to do)?
Cheers,
Loris