[slurm-users] slurm job_container/tmpfs

René Sitt sittr at hrz.uni-marburg.de
Tue Nov 21 13:35:22 UTC 2023


Hello Alain,

as an alternative to job_container/tmpfs, you may also try your luck 
with the 'auto_tmpdir' SPANK plugin: 
https://github.com/University-of-Delaware-IT-RCI/auto_tmpdir

We've been using using that on our small HPC cluster (Slurm 22.05) and 
it does what it's supposed to. One thing one has to remember is that it 
requires a recompile after every Slurm update.

Kind regards,
René Sitt

Am 21.11.23 um 14:07 schrieb Arsene Marian Alain:
>
> Thanks Sean. I’ve tried using slurm prolog/epilog scripts but without 
> any success. That's why I decided to look for other solutions and 
> job_container/tmpfs plugin seemed like a good alternative.
>
> *De:* slurm-users <slurm-users-bounces at lists.schedmd.com> *En nombre 
> de *Sean Mc Grath
> *Enviado el:* martes, 21 de noviembre de 2023 12:57
> *Para:* Slurm User Community List <slurm-users at lists.schedmd.com>
> *Asunto:* Re: [slurm-users] slurm job_container/tmpfs
>
> *ATENCIÓN*:Este correo electrónico se envió desde fuera de la UAH. No 
> haga clic en enlaces ni abra archivos adjuntos a menos que reconozca 
> al remitente y sepa que el contenido es seguro.
>
> Would a prolog script, https://slurm.schedmd.com/prolog_epilog.html, 
> do what you need? Sorry if you have already considered that and I 
> missed it.
>
> ---
>
> Sean McGrath
>
> Senior Systems Administrator, IT Services
>
> ------------------------------------------------------------------------
>
> *From:*slurm-users <slurm-users-bounces at lists.schedmd.com> on behalf 
> of Arsene Marian Alain <alain.arsene at uah.es>
> *Sent:* Tuesday 21 November 2023 09:58
> *To:* Slurm User Community List <slurm-users at lists.schedmd.com>
> *Subject:* Re: [slurm-users] slurm job_container/tmpfs
>
> Hello Brian,
>
> Thanks for your answer. With the job_container/tmpfs plugin I don't 
> really create the directory manually.
>
> I just give my Basepath=/scratch (a local directory for each node that 
> is already mounted with 1777 permissions) in job_container.conf. The 
> plugin automatically generates for each job a directory with the 
> "JOB_ID", for example: /scratch/1805
>
> The only problem is that directory 1805 is generated with root owner 
> and permissions 700. So the user who submitted the job cannot 
> write/read inside directory 1805.
>
> Is there a way for the owner of directory 1805 to be the user who 
> submitted the job and not root?
>
> *De:* slurm-users <slurm-users-bounces at lists.schedmd.com> *En nombre 
> de *Brian Andrus
> *Enviado el:* lunes, 20 de noviembre de 2023 23:29
> *Para:* slurm-users at lists.schedmd.com
> *Asunto:* Re: [slurm-users] slurm job_container/tmpfs
>
> *ATENCIÓN*:Este correo electrónico se envió desde fuera de la UAH. No 
> haga clic en enlaces ni abra archivos adjuntos a menos que reconozca 
> al remitente y sepa que el contenido es seguro.
>
> How do you 'manually create a directory'? That would be when the 
> ownership of root would be occurring. After creating it, you can 
> chown/chmod it as well.
>
> Brian Andrus
>
> On 11/18/2023 7:35 AM, Arsene Marian Alain wrote:
>
>     Dear slurm community,
>
>     I run slurm 21.08.1 under Rocky Linux 8.5 on my small HPC cluster
>     and am trying to configure job_container/tmpfs to manage the
>     temporary directories.
>
>     I have a shared nfs drive "/home" and a local "/scratch" (with
>     permissions 1777) on each node.
>
>     For each submitted job I manually create a directory with the
>     "JOB_ID.$USER" in the local "/scratch" which is where all the temp
>     files for the job will be generated. Now, I would like to do these
>     automatically (especially to remove the directory when the job
>     finishes or is canceled):
>
>     I added the following parameters in my /etc/slurm.conf:
>
>     JobContainerType=job_container/tmpfs
>
>     PrologFlags=contain
>
>     So, I have created the "job_container.conf" in the directory
>     "/etc/slurm"
>
>     with the following configuration:
>
>     AutoBasePath=false
>
>     BasePath=/scratch
>
>     Then, I replicated the changes to all nodes and restarted the
>     slurm daemons.
>
>     Finally, when I launch the job a directory with the "JOB_ID" is
>     created in the local "/scratch" of the compute node. The only
>     problem is that the owner of the directory is "root" and the user
>     who submitted the job doesn’t have read and write permissions to
>     that directory (other users do not either).
>
>     I would like that:
>
>     1) The name of the automatically created directory will be:
>     "JOB_ID.$USER"
>
>     2) The owner of the directory will be the user who submitted the
>     job, not "root".
>
>     Please, could someone help me?
>
>     Thanks a lot.
>
>     Best regards,
>
>     Alain
>
-- 
Dipl.-Chem. René Sitt
Hessisches Kompetenzzentrum für Hochleistungsrechnen
Philipps-Universität Marburg
Hans-Meerwein-Straße
35032 Marburg

Tel. +49 6421 28 23523
sittr at hrz.uni-marburg.de
www.hkhlr.de
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20231121/2ccc1821/attachment.htm>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: smime.p7s
Type: application/pkcs7-signature
Size: 4239 bytes
Desc: Kryptografische S/MIME-Signatur
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20231121/2ccc1821/attachment.bin>


More information about the slurm-users mailing list