[slurm-users] GPUs as resources which SLURM can control
Nicholas Yue
yue.nicholas at gmail.com
Wed Mar 20 23:05:57 UTC 2019
Hi,
I am new to SLURM.
I have access to a cluster where one of the node has 4 GPUs
We are running version SLURM 17.11.12
Is there some SBATCH token=value pair value I can use to submit jobs
(each of which has an application that is only able to utilize 1 GPU) so
that if I submit 6 copies, 4 copies will be dispatched and the 2 remaining
will be in a state e.g. PD, until a GPU frees up
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 396.44 Driver Version:
396.44 |
|-------------------------------+----------------------+----------------------+
| GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr.
ECC |
| Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute
M. |
|===============================+======================+======================|
| 0 Tesla P100-PCIE... On | 00000000:25:00.0 Off |
0 |
| N/A 29C P0 26W / 250W | 0MiB / 16280MiB | 0%
Default |
+-------------------------------+----------------------+----------------------+
| 1 Tesla P100-PCIE... On | 00000000:59:00.0 Off |
0 |
| N/A 26C P0 26W / 250W | 0MiB / 16280MiB | 0%
Default |
+-------------------------------+----------------------+----------------------+
| 2 Tesla P100-PCIE... On | 00000000:6D:00.0 Off |
0 |
| N/A 27C P0 26W / 250W | 0MiB / 16280MiB | 0%
Default |
+-------------------------------+----------------------+----------------------+
| 3 Tesla P100-PCIE... On | 00000000:99:00.0 Off |
0 |
| N/A 31C P0 26W / 250W | 0MiB / 16280MiB | 0%
Default |
+-------------------------------+----------------------+----------------------+
Cheers
--
Nicholas Yue
Graphics - Arnold, Alembic, RenderMan, OpenGL, HDF5
Custom Dev - C++ porting, OSX, Linux, Windows
http://au.linkedin.com/in/nicholasyue
https://vimeo.com/channels/naiadtools
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20190321/776fc287/attachment.html>
More information about the slurm-users
mailing list