[slurm-users] Reading which GPUs were assigned to which job
Stephan Roth
stephan.roth at ee.ethz.ch
Thu Apr 23 11:49:36 UTC 2020
Hello Manuel,
One way I know of is to use scontrol with the '-d' option:
scontrol -d show job=<JOBID>
Then the GPU is listed in a line like this:
Nodes=node01 CPU_IDs=14 Mem=10240 GRES=gpu(IDX:2)
If there are other ways to achieve the same, I'd be interested as well.
Best,
Stephan
-------------------------------------------------------------------
Stephan Roth | ISG.EE D-ITET ETH Zurich | http://www.isg.ee.ethz.ch
+4144 632 30 59 | ETF D 104 | Sternwartstrasse 7 | 8092 Zurich
-------------------------------------------------------------------
On 23.04.20 10:48, Holtgrewe, Manuel wrote:
> Dear all,
>
> is it possible to find out which GPU was assigned to which job through
> squeue or sacct?
>
> My motivation is as follows: some users write jobs with bad resource
> usage (e.g., 1h CPU to precompute, followed by 1h GPU to process, and so
> on). I don't care so much about CPUs at the moment as that's not the
> bottleneck at the moment but GPUs are.
>
> What is the best way to approach this?
>
> Best wishes,
>
> --
> Dr. Manuel Holtgrewe, Dipl.-Inform.
> Bioinformatician
> Core Unit Bioinformatics – CUBI
> Berlin Institute of Health / Max Delbrück Center for Molecular Medicine
> in the Helmholtz Association / Charité – Universitätsmedizin Berlin
>
> Visiting Address: Invalidenstr. 80, 3rd Floor, Room 03 028, 10117 Berlin
> Postal Address: Chariteplatz 1, 10117 Berlin
>
> E-Mail: manuel.holtgrewe at bihealth.de
> Phone: +49 30 450 543 607
> Fax: +49 30 450 7 543 901
> Web: cubi.bihealth.org www.bihealth.org www.mdc-berlin.de www.charite.de
More information about the slurm-users
mailing list