[slurm-users] slurm-users Digest, Vol 37, Issue 33

vero chaul verochaul at gmail.com
Wed Nov 18 12:01:07 UTC 2020


Baja

El El mié, 18 nov. 2020 a la(s) 09:00, <
slurm-users-request at lists.schedmd.com> escribió:

> Send slurm-users mailing list submissions to
>         slurm-users at lists.schedmd.com
>
> To subscribe or unsubscribe via the World Wide Web, visit
>         https://lists.schedmd.com/cgi-bin/mailman/listinfo/slurm-users
> or, via email, send a message with subject or body 'help' to
>         slurm-users-request at lists.schedmd.com
>
> You can reach the person managing the list at
>         slurm-users-owner at lists.schedmd.com
>
> When replying, please edit your Subject line so it is more specific
> than "Re: Contents of slurm-users digest..."
>
>
> Today's Topics:
>
>    1. Just one node getting job allocations (La Pulga Encuera y Rabiosa)
>    2. Slurm version 20.11.0 is now available (Tim Wickberg)
>    3. Sreport Query (navin srivastava)
>    4. Re: Sreport Query (navin srivastava)
>
>
> ----------------------------------------------------------------------
>
> Message: 1
> Date: Tue, 17 Nov 2020 10:50:17 -0500
> From: La Pulga Encuera y Rabiosa <lapulgaencuerayrabiosa at gmail.com>
> To: "slurm-users at lists.schedmd.com" <slurm-users at lists.schedmd.com>
> Subject: [slurm-users] Just one node getting job allocations
> Message-ID:
>         <
> CAP3Hg1G7msGsK8cFKCmAmmPoa45_HEKYUi18-xT7_uH-njSeBg at mail.gmail.com>
> Content-Type: text/plain; charset="utf-8"
>
> And if I try to run another job and all resources in that one node are all
> used then the job is pur to pending. I'm running srun getting pseudo
> terminals allocations to install some Spack packages. this node has 40
> cores (2 sockets @20 cores each). same specs and memory size for the other
> nodes. Any clues where is the issue or where to start looking?
>
> Thank you all.
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: <
> http://lists.schedmd.com/pipermail/slurm-users/attachments/20201117/9302134f/attachment-0001.htm
> >
>
> ------------------------------
>
> Message: 2
> Date: Tue, 17 Nov 2020 10:32:38 -0700
> From: Tim Wickberg <tim at schedmd.com>
> To: slurm-announce at schedmd.com, slurm-users at schedmd.com
> Subject: [slurm-users] Slurm version 20.11.0 is now available
> Message-ID: <312be1b3-bf71-cb3a-4a2e-2e56769318de at schedmd.com>
> Content-Type: text/plain; charset=utf-8; format=flowed
>
> After 9 months of development and testing we are pleased to announce the
> availability of Slurm version 20.11.0!
>
> Slurm 20.11 includes a number of new features including:
>
> - Overhaul of the job step management and launch code, alongside
> improved GPU task placement support.
>
> - A new "Interactive Step" mode of operation for salloc.
>
> - A new "scrontab" command that can be used to submit and manage
> periodically repeating jobs.
>
> - IPv6 support.
>
> - Changes to the reservation logic, with new options allowing users to
> delete reservations, allowing admins to skip the next occurance of a
> repeated reservation, and allowing for a job to be submitted and
> eligible to run within multiple reservations.
>
> - Dynamic Future Nodes - automatically associate a dynamically
> provisioned (or "cloud") node against a NodeName definition with
> matching hardware.
>
> - An experimental new RPC queuing mode for slurmctld to reduce thread
> contention on heavily loaded clusters.
>
> - SlurmDBD integration with the Slurm REST API.
>
> Please see the RELEASE_NOTES distributed alongside the source for
> further details.
>
> Thank you to all customers, partners, and community members who
> contributed to this release.
>
> As with past releases, the documentation available at
> https://slurm.schedmd.com has been updated to the 20.11 release. Past
> versions are available in the archive. This release also marks the end
> of support for the 19.05 release. The 20.02 release will remain
> supported up until the 21.08 release next August, but will not see as
> frequent updates, and bug-fixes will be targeted for the 20.11
> maintenance releases going forward.
>
> Slurm can be downloaded from https://www.schedmd.com/downloads.php .
>
> - Tim
>
> --
> Tim Wickberg
> Chief Technology Officer, SchedMD LLC
> Commercial Slurm Development and Support
>
>
>
> ------------------------------
>
> Message: 3
> Date: Wed, 18 Nov 2020 10:37:00 +0530
> From: navin srivastava <navin.altair at gmail.com>
> To: Slurm User Community List <slurm-users at lists.schedmd.com>
> Subject: [slurm-users] Sreport Query
> Message-ID:
>         <
> CAK8-jZAnoJ0A_q6isCFWjcsAEpht+FDctEZPiqkpEcahx0M7Mg at mail.gmail.com>
> Content-Type: text/plain; charset="utf-8"
>
> Dear All,
>
> Good Day!
>
> i am seeing one strange behaviour in my environment.
>
> we have 2 clusters in our environment one acting as a database server and
> have pointed the 2nd cluster to the same database.
>
> ------------------ ---------
>   hpc1  155.250.126.30         6817  8192         1
>                                                                   normal
>   hpc2  155.250.168.57         6817  8192         1
>                                                                   normal
>
> While generating the report I am able to generate for the local
> cluster(hpc1) without any issue  and it looks good. but from the second
> cluster data it always shows me 100% utilization from june onwards ,earlier
> data is fine.which is definitely wrong.
>
> sreport cluster utilization start=06/01/20 end=06/30/20 cluster=hpc2 -t
> percent | grep hpc2
> hpc2     100.00%    0.00%    0.00%    0.00%    0.00%     99.82%
>
> any suggestion what went wrong here. how to troubleshoot this issue.
>
> Regards
> Navin.
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: <
> http://lists.schedmd.com/pipermail/slurm-users/attachments/20201118/41c08789/attachment-0001.htm
> >
>
> ------------------------------
>
> Message: 4
> Date: Wed, 18 Nov 2020 12:11:22 +0530
> From: navin srivastava <navin.altair at gmail.com>
> To: Slurm User Community List <slurm-users at lists.schedmd.com>
> Subject: Re: [slurm-users] Sreport Query
> Message-ID:
>         <CAK8-jZBH2wYrxTfh5w60=24JtHbvHL=
> Agq762c0yH1YZCQmP2Q at mail.gmail.com>
> Content-Type: text/plain; charset="utf-8"
>
> is there a way to find the utilization per Node?
>
> Regards
> Navin.
>
> On Wed, Nov 18, 2020 at 10:37 AM navin srivastava <navin.altair at gmail.com>
> wrote:
>
> > Dear All,
> >
> > Good Day!
> >
> > i am seeing one strange behaviour in my environment.
> >
> > we have 2 clusters in our environment one acting as a database server and
> > have pointed the 2nd cluster to the same database.
> >
> > ------------------ ---------
> >   hpc1  155.250.126.30         6817  8192         1
> >
>  normal
> >   hpc2  155.250.168.57         6817  8192         1
> >
>  normal
> >
> > While generating the report I am able to generate for the local
> > cluster(hpc1) without any issue  and it looks good. but from the second
> > cluster data it always shows me 100% utilization from june onwards
> ,earlier
> > data is fine.which is definitely wrong.
> >
> > sreport cluster utilization start=06/01/20 end=06/30/20 cluster=hpc2 -t
> > percent | grep hpc2
> > hpc2     100.00%    0.00%    0.00%    0.00%    0.00%     99.82%
> >
> > any suggestion what went wrong here. how to troubleshoot this issue.
> >
> > Regards
> > Navin.
> >
> >
> >
> >
> >
> >
> >
> >
> >
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL: <
> http://lists.schedmd.com/pipermail/slurm-users/attachments/20201118/0ab04899/attachment-0001.htm
> >
>
> End of slurm-users Digest, Vol 37, Issue 33
> *******************************************
>
-- 
Veronica Chaul
+5411 3581-4041
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20201118/46e16b88/attachment.htm>


More information about the slurm-users mailing list