[slurm-users] srun problem -- Can't find an address, check slurm.conf

Scott Hazelhurst Scott.Hazelhurst at wits.ac.za
Tue Nov 13 01:24:53 MST 2018


Dear all

I still haven’t found the cause to the problem I raised last week where srun -w xx  runs for some nodes but not for others — thanks for the ideas.

One intriguing result I’ve had trying to pursue this which I thought I’d share in case it sparks some ideas. If I give the full path for srun, then it works


# show path
scott at cream-ce ~]$ which srun
/opt/exp_soft/bin/srun


# Node n37 is good (as are most of our nodes)
[scott at cream-ce ~]$ srun  -w n37 --pty bash
[scott at n37 ~]$ 


# Node n38 is not (and a few othrs)
scott at cream-ce ~]$ srun  -w n38 --pty bash
srun: error: fwd_tree_thread: can't find address for host n38, check slurm.conf
srun: error: Task launch for 20094.0 failed on node n38: Can't find an address, check slurm.conf
srun: error: Application launch failed: Can't find an address, check slurm.conf
srun: Job step aborted: Waiting up to 32 seconds for job step to finish.
srun: error: Timed out waiting for job step to complete

But if I  give the full path name — it works!

scott at cream-ce ~]$ /opt/exp_soft/slurm/bin/srun  -w n38 --pty bash
[scott at n38 ~]$ 


Scott



Scott


This communication is intended for the addressee only. It is confidential. If you have received this communication in error, please notify us immediately and destroy the original message. You may not copy or disseminate this communication without the permission of the University. Only authorised signatories are competent to enter into agreements on behalf of the University and recipients are thus advised that the content of this message may not be legally binding on the University and may contain the personal views and opinions of the author, which are not necessarily the views and opinions of The University of the Witwatersrand, Johannesburg. All agreements between the University and outsiders are subject to South African Law unless the University agrees in writing to the contrary.


More information about the slurm-users mailing list