[slurm-users] network/communication failure

Turner, Heath Hturner at eng.ua.edu
Mon May 21 09:46:58 MDT 2018


Got it! It was the firewall...

Thanks to all for all the suggestions.

Heath

Professor
Graduate Coordinator
Chemical and Biological Engineering
http://che.eng.ua.edu
 
University of Alabama
3448 SEC, Box 870203
Tuscaloosa, AL  35487
(205) 348-1733 (phone)
(205) 561-7450 (cell)
(205) 348-7558 (fax)
hturner at eng.ua.edu
http://turnerresearchgroup.ua.edu

-----Original Message-----
From: slurm-users [mailto:slurm-users-bounces at lists.schedmd.com] On Behalf Of Andy Riebs
Sent: Monday, May 21, 2018 10:22 AM
To: slurm-users at lists.schedmd.com
Subject: Re: [slurm-users] network/communication failure

Do you have a firewall running?

On 05/21/2018 11:05 AM, Turner, Heath wrote:
> If anyone has advice, I would really appreciate...
>
> I am running (just installed) slurm-11.17.6, with a master + 2 hosts.  It works locally on the master (controller + execution).  However, I cannot establish communication from master [triumph01] with the 2 hosts [triumph02,triumph03].  Here is some more info:
>
> 1. munge is running, and munge verification tests all pass.
> 2. system clocks are in sync on master/hosts.
> 3. identical slurm.conf files are on master/hosts.
> 4. configuration of resources (memory/cpus/etc) are correct and have been confirmed on all machines (all hardware is identical).
> 5. I have attached:
> 	a) slurm.conf
> 	b) log file from master slurmctld
> 	c) log file from host slurmd
>
> Any ideas about what to try next?
>
> Heath Turner
>
> Professor
> Graduate Coordinator
> Chemical and Biological Engineering
> http://che.eng.ua.edu
>   
> University of Alabama
> 3448 SEC, Box 870203
> Tuscaloosa, AL  35487
> (205) 348-1733 (phone)
> (205) 561-7450 (cell)
> (205) 348-7558 (fax)
> hturner at eng.ua.edu
> http://turnerresearchgroup.ua.edu
>

-- 
Andy Riebs
andy.riebs at hpe.com
Hewlett-Packard Enterprise
High Performance Computing Software Engineering
+1 404 648 9024
My opinions are not necessarily those of HPE
     May the source be with you!





More information about the slurm-users mailing list