[slurm-users] Strange problem with Slurm 17.11.0: "batch job complete failure"

Andy Riebs andy.riebs at hpe.com
Fri Dec 8 09:06:47 MST 2017


I've gathered more information, and I am probably having a fight with 
pam.  First, of note, this problem can be reproduced with a single node, 
single task job, such as

$ sbatch -N1 --reservation awr
#!/bin/bash
hostname
Submitted batch job 90436
$ sinfo -R
batch job complete f slurm     2017-12-08T15:34:37 node017
$

With SlurmdDebug=debug5, the only thing interesting in slurmd.log is

[2017-12-08T15:34:37.770] [90436.batch] Munge cryptographic signature 
plugin loaded
[2017-12-08T15:34:37.778] [90436.batch] error: pam_open_session: Cannot 
make/remove an entry for the specified session
[2017-12-08T15:34:37.779] [90436.batch] error: error in pam_setup
[2017-12-08T15:34:37.804] [90436.batch] error: job_manager exiting 
abnormally, rc = 4020
[2017-12-08T15:34:37.804] [90436.batch] job 90436 completed with 
slurm_rc = 4020, job_rc = 0

/etc/pam.d/slurm is defined as

auth            required        pam_localuser.so
auth            required        pam_shells.so
account         required        pam_unix.so
account         required        pam_access.so
session         required        pam_unix.so
session         required        pam_loginuid.so

/var/log/secure reports

Dec  8 15:34:37 node017 : pam_unix(slurm:session): open_session - error 
recovering username
Dec  8 15:34:37 node017 : pam_loginuid(slurm:session): unexpected 
response from failed conversation function
Dec  8 15:34:37 node017 : pam_loginuid(slurm:session): error recovering 
login user-name

The message "error recovering username" seems likely to be at the heart 
of the problem here. This worked just fine with Slurm 16.05.8, and I 
think it was also working with Slurm 17.11.0-0pre2.

Any thoughts about where I should go from here?

Andy

On 11/30/2017 08:40 AM, Andy Riebs wrote:
> We've just installed 17.11.0 on our 100+ node x86_64 cluster running 
> CentOS 7.4 this afternoon, and periodically see a single node (perhaps 
> the first node in an allocation?) get drained with the message "batch 
> job complete failure".
>
> On one node in question, slurmd.log reports
>
>     pam_unix(slurm:session): open_session - error recovering username
>     pam_loginuid(slurm:session): unexpected response from failed
>     conversation function 
>
> On another node drained for the same reason,
>
>     error: pam_open_session: Cannot make/remove an entry for the
>     specified session
>     error: error in pam_setup
>     error: job_manager exiting abnormally, rc = 4020
>     sending REQUEST_COMPLETE_BATCH_SCRIPT, error:4020 status 0
>
> slurmctld has logged
>
>     error: slurmd error running JobId=33 on node(s)=node048: Slurmd
>     could not execve job
>
>     drain_nodes: node node048 state set to DRAIN
>
> If anyone can shine some light on where I should start looking, I 
> shall be most obliged!
>
> Andy
>
> -- 
> Andy Riebs
> andy.riebs at hpe.com
> Hewlett-Packard Enterprise
> High Performance Computing Software Engineering
> +1 404 648 9024
> My opinions are not necessarily those of HPE
>      May the source be with you!

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20171208/5830fe56/attachment.html>


More information about the slurm-users mailing list