[slurm-users] How to apply for multiple GPU cards from different worker nodes?

Antony Cleave antony.cleave at gmail.com
Mon Apr 15 08:38:24 UTC 2019


Ask for 8 gpus on 2 nodes instead.

In your script just change the 16 to 8 and it should do what you want.

You are currently asking for 2 nodes with 16 gpu each as Gres resources are
per node.

Antony

On Mon, 15 Apr 2019, 09:08 Ran Du, <bella.ran.du at gmail.com> wrote:

> Dear all,
>
>      Does anyone know how to set #SBATCH options to get multiple GPU cards
> from different worker nodes?
>
>      One of our users would like to apply for 16 NVIDIA V100 cards for his
> job, and  there are 8 GPU cards on each worker node, I have tried the
> following #SBATCH options:
>
>       #SBATCH --partition=gpu
>       #SBATCH --qos=normal
>       #SBATCH --account=u07
>       #SBATCH --job-name=cross
>       #SBATCH --nodes=2
>       #SBATCH --mem-per-cpu=1024
>       #SBATCH --output=test.32^4.16gpu.log
>       #SBATCH --gres=gpu:v100:16
>
>       but got the sbatch error message :
>       sbatch: error: Batch job submission failed: Requested node
> configuration is not available
>
>       And I found a similar question on stack overflow:
>
> https://stackoverflow.com/questions/45200926/how-to-access-to-gpus-on-different-nodes-in-a-cluster-with-slurm
>
>       And it is said that multiple GPU cards allocation on different
> worker nodes are not available, the post is in 2017, is it still true at
> present?
>
>       Thanks a lot for your help.
>
> Best regards,
> Ran
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20190415/a0c92aa4/attachment.html>


More information about the slurm-users mailing list