[slurm-users] Slurm 20.02.3 error: CPUs=1 match no Sockets, Sockets*CoresPerSocket or Sockets*CoresPerSocket*ThreadsPerCore. Resetting CPUs.

Jeffrey T Frey frey at udel.edu
Tue Jun 16 13:48:12 UTC 2020


If you check the source up on Github, that's more of a warning produced when you didn't specify a CPU count and it's going to calculate from the socket-core-thread numbers (src/common/read_config.c):



		/* Node boards are factored into sockets */
		if ((n->cpus != n->sockets) &&
		    (n->cpus != n->sockets * n->cores) &&
		    (n->cpus != n->sockets * n->cores * n->threads)) {
			error("NodeNames=%s CPUs=%d match no Sockets, Sockets*CoresPerSocket or Sockets*CoresPerSocket*ThreadsPerCore. Resetting CPUs.",
			      n->nodenames, n->cpus);
			n->cpus = n->sockets * n->cores * n->threads;
		}


This behavior is present beginning in 18.x releases; in 17.x and earlier the inferred n->cpus was done quietly.


> On Jun 16, 2020, at 04:12 , Ole Holm Nielsen <Ole.H.Nielsen at fysik.dtu.dk> wrote:
> 
> Today we upgraded the controller node from 19.05 to 20.02.3, and immediately all Slurm commands (on the controller node) give error messages for all partitions:
> 
> # sinfo --version
> sinfo: error: NodeNames=a[001-140] CPUs=1 match no Sockets, Sockets*CoresPerSocket or Sockets*CoresPerSocket*ThreadsPerCore. Resetting CPUs.
> (lines deleted)
> slurm 20.02.3
> 
> In slurm.conf we have defined NodeName like:
> 
> NodeName=a[001-140] Weight=10001 Boards=1 SocketsPerBoard=2 CoresPerSocket=4 ThreadsPerCore=1 ...
> 
> According to the slurm.conf manual the CPUs should then be calculated automatically:
> 
> "If CPUs is omitted, its default will be set equal to the product of Boards, Sockets, CoresPerSocket, and ThreadsPerCore."
> 
> Has anyone else seen this error with Slurm 20.02?
> 
> I wonder if there is a problem with specifying SocketsPerBoard in stead of Sockets?  The slurm.conf manual doesn't seem to prefer one over the other.
> 
> I've opened a bug https://bugs.schedmd.com/show_bug.cgi?id=9241
> 
> Thanks,
> Ole
> 
> 

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20200616/f79a6371/attachment.htm>


More information about the slurm-users mailing list