We are pleased to announce the availability of Slurm release candidate
25.05.0rc1.
To highlight some new features coming in 25.05:
- Support for defining multiple topology configurations, and varying
them by partition.
- Support for tracking and allocating hierarchical resources.
- Dynamic nodes can be dynamically added to the topology.
- topology/block - Allow for gaps in the block layout.
- Support for encrypting all network communication with TLS.
- jobcomp/kafka - Optionally send job info at job start as well as job end.
- Support an OR operator in --license requests.
- switch/hpe_slingshot - Support for > 252 ranks per node.
- switch/hpe_slingshot - Support mTLS authentication to the fabric manager.
- sacctmgr - Add support for dumping and loading QOSes.
- srun - Add new --wait-for-children option to keep the step running
until all launched processes have been launched (cgroup/v2 only).
- slurmrestd - Add new endpoint for creating reservations.
This is the first release candidate of the upcoming 25.05 release
series, and represents the end of development for this release, and a
finalization of the RPC and state file formats.
If any issues are identified with this release candidate, please report
them through https://bugs.schedmd.com against the 25.05.x version and we
will address them before the first production 25.05.0 release is made.
Please note that the release candidates are not intended for production use.
A preview of the updated documentation can be found at
https://slurm.schedmd.com/archive/slurm-master/ .
Slurm can be downloaded from https://www.schedmd.com/download-slurm/.
The changelog for 25.05.0rc1 can be found here:
https://github.com/SchedMD/slurm/blob/master/CHANGELOG/slurm-25.05.md#chang…
--
Marshall Garey
Release Management, Support, and Development
SchedMD LLC - Commercial Slurm Development and Support
Slurm versions 24.11.5, 24.05.8, and 23.11.11 are now available and
include a fix for a recently discovered security issue.
SchedMD customers were informed on April 23rd and provided a patch on
request; this process is documented in our security policy. [1]
A mistake with permission handling for Coordinators within Slurm's
accounting system can allow a Coordinator to promote a user to
Administrator. (CVE-2025-43904)
Thank you to Sekou Diakite (HPE) for reporting this.
Downloads are available at https://www.schedmd.com/downloads.php .
Release notes follow below.
- Tim
[1] https://www.schedmd.com/security-policy/
--
Tim Wickberg
Chief Technology Officer, SchedMD LLC
Commercial Slurm Development and Support
> * Changes in Slurm 24.11.5
> ==========================
> -- Return error to scontrol reboot on bad nodelists.
> -- slurmrestd - Report an error when QOS resolution fails for v0.0.40
> endpoints.
> -- slurmrestd - Report an error when QOS resolution fails for v0.0.41
> endpoints.
> -- slurmrestd - Report an error when QOS resolution fails for v0.0.42
> endpoints.
> -- data_parser/v0.0.42 - Added +inline_enums flag which modifies the
> output when generating OpenAPI specification. It causes enum arrays to not
> be defined in their own schema with references ($ref) to them. Instead they
> will be dumped inline.
> -- Fix binding error with tres-bind map/mask on partial node allocations.
> -- Fix stepmgr enabled steps being able to request features.
> -- Reject step creation if requested feature is not available in job.
> -- slurmd - Restrict listening for new incoming RPC requests further into
> startup.
> -- slurmd - Avoid auth/slurm related hangs of CLI commands during startup
> and shutdown.
> -- slurmctld - Restrict processing new incoming RPC requests further into
> startup. Stop processing requests sooner during shutdown.
> -- slurmcltd - Avoid auth/slurm related hangs of CLI commands during
> startup and shutdown.
> -- slurmctld: Avoid race condition during shutdown or reconfigure that
> could result in a crash due delayed processing of a connection while
> plugins are unloaded.
> -- Fix small memleak when getting the job list from the database.
> -- Fix incorrect printing of % escape characters when printing stdio
> fields for jobs.
> -- Fix padding parsing when printing stdio fields for jobs.
> -- Fix printing %A array job id when expanding patterns.
> -- Fix reservations causing jobs to be held for Bad Constraints
> -- switch/hpe_slingshot - Prevent potential segfault on failed curl
> request to the fabric manager.
> -- Fix printing incorrect array job id when expanding stdio file names.
> The %A will now be substituted by the correct value.
> -- Fix printing incorrect array job id when expanding stdio file names.
> The %A will now be substituted by the correct value.
> -- switch/hpe_slingshot - Fix vni range not updating on slurmctld restart
> or reconfigre.
> -- Fix steps not being created when using certain combinations of -c and
> -n inferior to the jobs requested resources, when using stepmgr and nodes
> are configured with CPUs == Sockets*CoresPerSocket.
> -- Permit configuring the number of retry attempts to destroy CXI service
> via the new destroy_retries SwitchParameter.
> -- Do not reset memory.high and memory.swap.max in slurmd startup or
> reconfigure as we are never really touching this in slurmd.
> -- Fix reconfigure failure of slurmd when it has been started manually and
> the CoreSpecLimits have been removed from slurm.conf.
> -- Set or reset CoreSpec limits when slurmd is reconfigured and it was
> started with systemd.
> -- switch/hpe-slingshot - Make sure the slurmctld can free step VNIs after
> the controller restarts or reconfigures while the job is running.
> -- Fix backup slurmctld failure on 2nd takeover.
> -- Testsuite - fix python test 130_2.
> -- Fix security issue where a coordinator could add a user with elevated
> privileges. CVE-2025-43904.
> * Changes in Slurm 24.05.8
> ==========================
> -- Testsuite - fix python test 130_2.
> -- Fix security issue where a coordinator could add a user with elevated
> privileges. CVE-2025-43904.
> * Changes in Slurm 23.11.11
> ===========================
> -- Fixed a job requeuing issue that merged job entries into the same SLUID
> when all nodes in a job failed simultaneously.
> -- Add ABORT_ON_FATAL environment variable to capture a backtrace from any
> fatal() message.
> -- Testsuite - fix python test 130_2.
> -- Fix security issue where a coordinator could add a user with elevated
> privileges. CVE-2025-43904.
We are pleased to announce the availability of Slurm version 24.11.4.
This release fixes a variety of major to minor severity bugs. Some edge
cases that caused jobs to pend forever are fixed. Notable stability
issues that are fixed include:
* slurmctld crashing upon receiving a certain heterogeneous job submission.
* slurmd crashing after a communications failure with a slurmstepd.
* A variety of race conditions related to receiving and processing
connections, including one that resulted in the slurmd ignoring new RPC
connections.
Downloads are available at https://www.schedmd.com/downloads.php .
--
Marshall Garey
Release Management, Support, and Development
SchedMD LLC - Commercial Slurm Development and Support
> -- slurmctld,slurmrestd - Avoid possible race condition that could have caused
> process to crash when listener socket was closed while accepting a new
> connection.
> -- slurmrestd - Avoid race condition that could have resulted in address
> logged for a UNIX socket to be incorrect.
> -- slurmrestd - Fix parameters in OpenAPI specification for the following
> endpoints to have "job_id" field:
> GET /slurm/v0.0.40/jobs/state/
> GET /slurm/v0.0.41/jobs/state/
> GET /slurm/v0.0.42/jobs/state/
> GET /slurm/v0.0.43/jobs/state/
> -- slurmd - Fix tracking of thread counts that could cause incoming
> connections to be ignored after burst of simultaneous incoming connections
> that trigger delayed response logic.
> -- Stepmgr - Avoid unnecessary SRUN_TIMEOUT forwarding to stepmgr.
> -- Fix jobs being scheduled on higher weighted powered down nodes.
> -- Fix how backfill scheduler filters nodes from the available nodes based on
> exclusive user and mcs_label requirements.
> -- acct_gather_energy/{gpu,ipmi} - Fix potential energy consumption adjustment
> calculation underflow.
> -- acct_gather_energy/ipmi - Fix regression introduced in 24.05.5 (which
> introduced the new way of preserving energy measurements through slurmd
> restarts) when EnergyIPMICalcAdjustment=yes.
> -- Prevent slurmctld deadlock in the assoc mgr.
> -- Fix memory leak when RestrictedCoresPerGPU is enabled.
> -- Fix preemptor jobs not entering execution due to wrong calculation of
> accounting policy limits.
> -- Fix certain job requests that were incorrectly denied with node
> configuration unavailable error.
> -- slurmd - Avoid crash due when slurmd has a communications failure with
> slurmstepd.
> -- Fix memory leak when parsing yaml input.
> -- Prevent slurmctld from showing error message about PreemptMode=GANG being a
> cluster-wide option for `scontrol update part` calls that don't attempt to
> modify partition PreemptMode.
> -- Fix setting GANG preemption on partition when updating PreemptMode with
> scontrol.
> -- Fix CoreSpec and MemSpec limits not being removed from previously
> configured slurmd.
> -- Avoid race condition that could lead to a deadlock when slurmd, slurmstepd,
> slurmctld, slurmrestd or sackd have a fatal event.
> -- Fix jobs using --ntasks-per-node and --mem keep pending forever when the
> requested mem divided by the number of cpus will surpass the configured
> MaxMemPerCPU.
> -- slurmd - Fix address logged upon new incoming RPC connection from "INVALID"
> to IP address.
> -- Fix memory leak when retrieving reservations. This affects scontrol, sinfo,
> sview, and the following slurmrestd endpoints:
> 'GET /slurm/{any_data_parser}/reservation/{reservation_name}'
> 'GET /slurm/{any_data_parser}/reservations'
> -- Log warning instead of debuflags=conmgr gated log when deferring new
> incoming connections when number of active connections exceed
> conmgr_max_connections.
> -- Avoid race condition that could result in worker thread pool not activating
> all threads at once after a reconfigure resulting in lower utilization of
> available CPU threads until enough internal activity wakes up all threads
> in the worker pool.
> -- Avoid theoretical race condition that could result in new incoming RPC
> socket connections being ignored after reconfigure.
> -- slurmd - Avoid race condition that could result in a state where new
> incoming RPC connections will always be ignored.
> -- Add ReconfigFlags=KeepNodeStateFuture to restore saved FUTURE node state on
> restart and reconfig instead of reverting to FUTURE state. This will be
> made the default in 25.05.
> -- Fix case where hetjob submit would cause slurmctld to crash.
> -- Fix jobs using --cpus-per-gpu and --mem keep pending forever when the
> requested mem divided by the number of cpus will surpass the configured
> MaxMemPerCPU.
> -- Enforce that jobs using --mem and several --*-per-* options do not violate
> the MaxMemPerCPU in place.
> -- slurmctld - Fix use-cases of jobs incorrectly pending held when --prefer
> features are not initially satisfied.
> -- slurmctld - Fix jobs incorrectly held when --prefer not satisfied in some
> use-cases.
> -- Ensure RestrictedCoresPerGPU and CoreSpecCount don't overlap.
We are pleased to announce the availability of Slurm version 24.05.7.
This release fixes some stability issues in 24.05, including a crash in
slurmctld after updating a reservation with an empty nodelist.
Downloads are available at https://www.schedmd.com/downloads.php .
--
Marshall Garey
Release Management, Support, and Development
SchedMD LLC - Commercial Slurm Development and Support
> * Changes in Slurm 24.05.7
> ==========================
> -- Fix slurmctld crash when after updating a reservation with an empty
> nodelist. The crash could occur after restarting slurmctld, or if
> downing/draining a node in the reservation with the REPLACE or REPLACE_DOWN
> flag.
> -- Fix jobs being scheduled on higher weighted powered down
> nodes.
> -- Fix memory leak when RestrictedCoresPerGPU is enabled.
> -- Prevent slurmctld deadlock in the assoc mgr.
We are pleased to announce the availability of Slurm version 24.11.3.
24.11.3 fixes the database cluster ID generation not being random, a
regression in which slurmd -G gave no output, a long-standing crash in
slurmctld after updating a reservation with an empty nodelist, and some
other minor to moderate bugs.
Downloads are available at https://www.schedmd.com/downloads.php .
--
Marshall Garey
Release Management, Support, and Development
SchedMD LLC - Commercial Slurm Development and Support
> * Changes in Slurm 24.11.3
> ==========================
> -- Fix race condition in slurmrestd that resulted in "Requested
> data_parser plugin does not support OpenAPI plugin" error being returned
> for valid endpoints.
> -- If multiple partitions are requested, set the SLURM_JOB_PARTITION
> output environment variable to the partition in which the job is running
> for salloc and srun in order to match the documentation and the behavior of
> sbatch.
> -- Fix regression where slurmd -G gives no output.
> -- Don't print misleading errors for stepmgr enabled steps.
> -- slurmrestd - Avoid connection to slurmdbd for the following
> endpoints:
> GET /slurm/v0.0.41/jobs
> GET /slurm/v0.0.41/job/{job_id}
> -- slurmrestd - Avoid connection to slurmdbd for the following
> endpoints:
> GET /slurm/v0.0.40/jobs
> GET /slurm/v0.0.40/job/{job_id}
> -- Significantly increase entropy of clusterid portion of the
> sluid by seeding the random number generator
> -- Avoid changing process name to "watch" from original daemon name.
> This could potentially breaking some monitoring scripts.
> -- Avoid slurmctld being killed by SIGALRM due to race condition
> at startup.
> -- Fix slurmctld crash when after updating a reservation with an empty
> nodelist. The crash could occur after restarting slurmctld, or if
> downing/draining a node in the reservation with the REPLACE or REPLACE_DOWN
> flag.
> -- Fix race between task/cgroup cpuset and jobacctgather/cgroup.
> The first was removing the pid from task_X cgroup directory causing
> memory limits to not be applied.
> -- srun - Fixed wrongly constructed SLURM_CPU_BIND env variable
> that could get propagated to downward srun calls in certain mpi
> environments, causing launch failures.
> -- slurmrestd - Fix possible memory leak when parsing arrays with
> data_parser/v0.0.40.
> -- slurmrestd - Fix possible memory leak when parsing arrays with
> data_parser/v0.0.41.
> -- slurmrestd - Fix possible memory leak when parsing arrays with
> data_parser/v0.0.42.
We are pleased to announce the availability of Slurm versions 24.11.2
and 24.05.6.
24.11.2 fixes a variety of minor to major bugs. Fixed regressions
include loading non-default QOS on pending jobs from pre-24.11 state,
pending jobs displaying QOS=(null) when not explicitly requesting a QOS,
running jobs that requested multiple partitions potentially having an
incorrect partition when slurmctld is restarted, and burst_buffer.lua
failing if slurm.conf is in a non-standard location. This release also
fixes a few crashes in slurmctld: crashing when a job that can preempt
requests --test-only, crasing when the scheduler evaluates a job on
nodes with suspended jobs, and crashing due to a long-standing bug
causing a job record without job_resrcs.
24.05.6 fixes sattach with auth/slurm, a slurmrestd crash when using
data_parser/v0.0.40, a slurmctld crash when using job suspension, a
performance regression for RPCs with large amounts of data, and some
other moderate severity bugs.
Downloads are available at https://www.schedmd.com/downloads.php .
--
Marshall Garey
Release Management, Support, and Development
SchedMD LLC - Commercial Slurm Development and Support
> * Changes in Slurm 24.11.2
> ==========================
> -- Fix segfault when submitting --test-only jobs that can preempt.
> -- Fix regression introduced in 23.11 that prevented the following
> flags from being added to a reservation on an update:
> DAILY, HOURLY, WEEKLY, WEEKDAY, and WEEKEND.
> -- Fix crash and issues evaluating job's suitability for running in
> nodes with already suspended job(s) there.
> -- Slurmctld will ensure that healthy nodes are not reported as
> UnavailableNodes in job reason codes.
> -- Fix handling of jobs submitted to a current reservation with
> flags OVERLAP,FLEX or OVERLAP,ANY_NODES when it overlaps nodes with a
> future maintenance reservation. When a job submission had a time limit that
> overlapped with the future maintenance reservation, it was rejected. Now
> the job is accepted but stays pending with the reason "ReqNodeNotAvail,
> Reserved for maintenance".
> -- pam_slurm_adopt - avoid errors when explicitly setting
> some arguments to the default value.
> -- Fix qos preemption with PreemptMode=SUSPEND
> -- slurmdbd - When changing a user's name update lineage
> at the same time.
> -- Fix regression in 24.11 in which burst_buffer.lua does not
> inherit the SLURM_CONF environment variable from slurmctld and fails to run
> if slurm.conf is in a non-standard location.
> -- Fix memory leak in slurmctld if select/linear and the
> PreemptParameters=reclaim_licenses options are both set in slurm.conf.
> Regression in 24.11.1.
> -- Fix running jobs, that requested multiple partitions, from
> potentially being set to the wrong partition on restart.
> -- switch/hpe_slingshot - Fix compatibility with newer cxi
> drivers, specifically when specifying disable_rdzv_get.
> -- Add ABORT_ON_FATAL environment variable to capture a backtrace
> from any fatal() message.
> -- Fix printing invalid address in rate limiting log statement.
> -- sched/backfill - Fix node state PLANNED not being cleared from
> fully allocated nodes during a backfill cycle.
> -- select/cons_tres - Fix future planning of jobs with bf_licenses.
> -- Prevent redundant "on_data returned rc: Rate limit exceeded,
> please retry momentarily" error message from being printed in
> slurmctld logs.
> -- Fix loading non-default QOS on pending jobs from pre-24.11 state.
> -- Fix pending jobs displaying QOS=(null) when not explicitly
> requesting a QOS.
> -- Fix segfault issue from job record with no job_resrcs
> -- Fix failing sacctmgr delete/modify/show account operations
> with where clauses.
> -- Fix regression in 24.11 in which Slurm daemons started catching
> several SIGTSTP, SIGTTIN and SIGUSR1 signals and ignored them, while before
> they were not ignoring them. This also caused slurmctld to not being
> able to shutdown after a SIGTSTP because slurmscriptd caught the signal
> and stopped while slurmctld ignored it. Unify and fix these situations and
> get back to the previous behavior for these signals.
> -- Document that SIGQUIT is no longer ignored by slurmctld,
> slurmdbd, and slurmd in 24.11. As of 24.11.0rc1, SIGQUIT is identical to
> SIGINT and SIGTERM for these daemons, but this change was not documented.
> -- Fix not considering nodes marked for reboot without ASAP
> in the scheduler.
> -- Remove the boot^ state on unexpected node reboot after
> return to service.
> -- Do not allow new jobs to start on a node which is being rebooted
> with the flag nextstate=resume.
> -- Prevent lower priority job running after cancelling an ASAP reboot.
> -- Fix srun jobs starting on nextstate=resume rebooting nodes.
>
> * Changes in Slurm 24.05.6
> ==========================
> -- data_parser/v0.0.40 - Prevent a segfault in the slurmrestd when
> dumping data with v0.0.40+complex data parser.
> -- Fix sattach when using auth/slurm.
> -- scrun - Add support '--all' argument for kill subcommand.
> -- Fix performance regression while packing larger RPCs.
> -- Fix crash and issues evaluating job's suitability for running in
> nodes with already suspended job(s) there.
> -- Fixed a job requeuing issue that merged job entries into the
> same SLUID when all nodes in a job failed simultaneously.
> -- switch/hpe_slingshot - Fix compatibility with newer cxi
> drivers, specifically when specifying disable_rdzv_get.
> -- Add ABORT_ON_FATAL environment variable to capture a backtrace
> from any fatal() message.
We are pleased to announce the availability of Slurm version 24.11.1.
This fixes a few possible crashes of the slurmctld and slurmrestd; a
regression in 24.11 which caused file transfers to a job with sbcast to
not join the job container namespace; mpi apps using Intel OPA, PSM2 and
OMPI 5.x when ran through srun; and various minor to moderate bugs.
Downloads are available at https://www.schedmd.com/downloads.php .
--
Marshall Garey
Release Management, Support, and Development
SchedMD LLC - Commercial Slurm Development and Support
> * Changes in Slurm 24.11.1
> ==========================
> -- With client commands MIN_MEMORY will show mem_per_tres if specified.
> -- Fix errno message about bad constraint
> -- slurmctld - Fix crash and possible split brain issue if the
> backup controller handles an scontrol reconfigure while in control
> before the primary resumes operation.
> -- Fix stepmgr not getting dynamic node addrs from the controller
> -- stepmgr - avoid "Unexpected missing socket" errors.
> -- Fix `scontrol show steps` with dynamic stepmgr
> -- Deny jobs using the "R:" option of --signal if PreemptMode=OFF
> globally.
> -- Force jobs using the "R:" option of --signal to be preemptable
> by requeue or cancel only. If PreemptMode on the partition or QOS is off
> or suspend, the job will default to using PreemptMode=cancel.
> -- If --mem-per-cpu exceeds MaxMemPerCPU, the number of cpus per
> task will always be increased even if --cpus-per-task was specified. This
> is needed to ensure each task gets the expected amount of memory.
> -- Fix compilation issue on OpenSUSE Leap 15
> -- Fix jobs using more nodes than needed when not using -N
> -- Fix issue with allocation being allocated less resources
> than needed when using --gres-flags=enforce-binding.
> -- select/cons_tres - Fix errors with MaxCpusPerSocket partition
> limit. Used cpus/cores weren't counted properly, nor limiting free ones
> to avail, when the socket was partially allocated, or the job request
> went beyond this limit.
> -- Fix issue when jobs were preempted for licenses even if there
> were enough licenses available.
> -- Fix srun ntasks calculation inside an allocation when nodes are
> requested using a min-max range.
> -- Print correct number of digits for TmpDisk in sdiag.
> -- Fix a regression in 24.11 which caused file transfers to a job
> with sbcast to not join the job container namespace.
> -- data_parser/v0.0.40 - Prevent a segfault in the slurmrestd when
> dumping data with v0.0.40+complex data parser.
> -- Remove logic to force lowercase GRES names.
> -- data_parser/v0.0.42 - Prevent the association id from always
> being dumped as NULL when parsing in complex mode. Instead it will now
> dump the id. This affects the following endpoints:
> GET slurmdb/v0.0.42/association
> GET slurmdb/v0.0.42/associations
> GET slurmdb/v0.0.42/config
> -- Fixed a job requeuing issue that merged job entries into the
> same SLUID when all nodes in a job failed simultaneously.
> -- When a job completes, try to give idle nodes to reservations with
> the REPLACE flag before allowing them to be allocated to jobs.
> -- Avoid expensive lookup of all associations when dumping or
> parsing for v0.0.42 endpoints.
> -- Avoid expensive lookup of all associations when dumping or
> parsing for v0.0.41 endpoints.
> -- Avoid expensive lookup of all associations when dumping or
> parsing for v0.0.40 endpoints.
> -- Fix segfault when testing jobs against nodes with invalid gres.
> -- Fix performance regression while packing larger RPCs.
> -- Document the new mcs/label plugin.
> -- job_container/tmpfs - Fix Xauthoirty file being created
> outside the container when EntireStepInNS is enabled.
> -- job_container/tmpfs - Fix spank_task_post_fork not always
> running in the container when EntireStepInNS is enabled.
> -- Fix a job potentially getting stuck in CG on permissions
> errors while setting up X11 forwarding.
> -- Fix error on X11 shutdown if Xauthority file was not created.
> -- slurmctld - Fix memory or fd leak if an RPC is recieved that
> is not registered for processing.
> -- Inject OMPI_MCA_orte_precondition_transports when using PMIx. This fixes
> mpi apps using Intel OPA, PSM2 and OMPI 5.x when ran through srun.
> -- Don't skip the first partition_job_depth jobs per partition.
> -- Fix gres allocation issue after controller restart.
> -- Fix issue where jobs requesting cpus-per-gpu hang in queue.
> -- switch/hpe_slingshot - Treat HTTP status forbidden the same as
> unauthorized, allowing for a graceful retry attempt.
We are pleased to announce the availability of Slurm version 24.05.5.
This release fixes a few potential crashes, several stepmgr bugs,
compatibility for sstat and sattach with newer version steps, and some
other minor bugs.
Downloads are available at https://www.schedmd.com/downloads.php .
--
Marshall Garey
Release Management, Support, and Development
SchedMD LLC - Commercial Slurm Development and Support
> * Changes in Slurm 24.05.5
> ==========================
> -- Fix issue signaling cron jobs resulting in unintended requeues.
> -- Fix slurmctld memory leak in implementation of HealthCheckNodeState=CYCLE.
> -- job_container/tmpfs - Fix SLURM_CONF env variable not being properly set.
> -- sched/backfill - Fix job's time_limit being overwritten by time_min for job
> arrays in some situations.
> -- RoutePart - fix segfault from incorrect memory allocation when node doesn't
> exist in any partition.
> -- slurmctld - Fix crash when a job is evaluated for a reservation after
> removal of a dynamic node.
> -- gpu/nvml - Attempt loading libnvidia-ml.so.1 as a fallback for failure in
> loading libnvidia-ml.so.
> -- slurmrestd - Fix populating non-required object fields of objects as '{}' in
> JSON/YAML instead of 'null' causing compiled OpenAPI clients to reject
> the response to 'GET /slurm/v0.0.40/jobs' due to validation failure of
> '.jobs[].job_resources'.
> -- Fix sstat/sattach protocol errors for steps on higher version slurmd's
> (regressions since 20.11.0rc1 and 16.05.1rc1 respectively).
> -- slurmd - Avoid a crash when starting slurmd version 24.05 with
> SlurmdSpoolDir files that have been upgraded to a newer major version of
> Slurm. Log warnings instead.
> -- Fix race condition in stepmgr step completion handling.
> -- Fix slurmctld segfault with stepmgr and MpiParams when running a job array.
> -- Fix requeued jobs keeping their priority until the decay thread happens.
> -- slurmctld - Fix crash and possible split brain issue if the
> backup controller handles an scontrol reconfigure while in control
> before the primary resumes operation.
> -- Fix stepmgr not getting dynamic node addrs from the controller
> -- stepmgr - avoid "Unexpected missing socket" errors.
> -- Fix `scontrol show steps` with dynamic stepmgr
> -- Support IPv6 in configless mode
We are pleased to announce the availability of the Slurm 24.11 release.
To highlight some new features in 24.11:
- New gpu/nvidia plugin. This does not rely on any NVIDIA libraries, and
will
build by default on all systems. It supports basic GPU detection and
management, but cannot currently identify GPU-to-GPU links, or provide
usage data as these are not exposed by the kernel driver.
- Add autodetected GPUs to the output from "slurmd -C".
- Added new QOS-based reports to "sreport".
- Revamped network I/O with the "conmgr" thread-pool model.
- Added new "hostlist function" syntax for management commands and
configuration files.
- switch/hpe_slingshot - Added support for hardware collectives setup
through
the fabric manager. (Requires SlurmctldParameters=enable_stepmgr)
- Added SchedulerParameters=bf_allow_magnetic_slot configuration option to
allow backfill planning for magnetic reservations.
- Added new "scontrol listjobs" and "liststeps" commands to complement
"listpids", and provide --json/--yaml output for all three subcommands.
- Allow jobs to be submitted against multiple QOSes.
- Added new experimental "oracle" backfill scheduling support, which permits
jobs to be delayed if the oracle function determines the reduced
fragmentation of the network topology is sufficiently advantageous.
- Improved responsiveness of the controller when jobs are requeued by
replacing the "db_index" identifier with a slurmctld-generated unique
identifier. ("SLUID")
- New options to job_container/tmpfs to permit site-specific scripts to
modify the namespace before user steps are launched, and to ensure all
steps are completely captured within that namespace.
The Slurm documentation has also been updated to the 24.11 release.
(Older versions can be found in the archive, linked from the main
documentation page.)
Slurm can be downloaded from https://www.schedmd.com/download-slurm/ .
- Tim
--
Tim Wickberg
Chief Technology Officer, SchedMD LLC
Commercial Slurm Development and Support
We are pleased to announce the availability of Slurm release candidate
24.11.0rc1.
To highlight some new features coming in 24.11:
- New gpu/nvidia plugin. This does not rely on any NVIDIA libraries, and
will build by default on all systems. It supports basic GPU detection
and management, but cannot currently identify GPU-to-GPU links, or
provide usage data as these are not exposed by the kernel driver.
- Add autodetected GPUs to the output from "slurmd -C".
- Added new QOS-based reports to "sreport".
- Revamped network I/O with the "conmgr" thread-pool model.
- Added new "hostlist function" syntax for management commands and
configuration files.
- switch/hpe_slingshot - Added support for hardware collectives setup
through the fabric manager. (Requires SlurmctldParameters=enable_stepmgr)
- Added SchedulerParameters=bf_allow_magnetic_slot configuration option
to allow backfill planning for magnetic reservations.
- Added new "scontrol listjobs" and "liststeps" commands to complement
"listpids", and provide --json/--yaml output for all three subcommands.
- Allow jobs to be submitted against multiple QOSes.
- Added new experimental "oracle" backfill scheduling support, which
permits jobs to be delayed if the oracle function determines the reduced
fragmentation of the network topology is sufficiently advantageous.
- Improved responsiveness of the controller when jobs are requeued by
replacing the "db_index" identifier with a slurmctld-generated unique
identifier. ("SLUID")
- New options to job_container/tmpfs to permit site-specific scripts to
modify the namespace before user steps are launched, and to ensure all
steps are completely captured within that namespace.
This is the first release candidate of the upcoming 24.11 release
series, and represents the end of development for this release, and a
finalization of the RPC and state file formats.
If any issues are identified with this release candidate, please report
them through https://bugs.schedmd.com against the 24.11.x version and we
will address them before the first production 24.11.0 release is made.
Please note that the release candidates are not intended for production use.
A preview of the updated documentation can be found at
https://slurm.schedmd.com/archive/slurm-master/ .
Slurm can be downloaded from https://www.schedmd.com/downloads.php .
--
Marshall Garey
Release Management, Support, and Development
SchedMD LLC - Commercial Slurm Development and Support
Slurm version 24.05.4 is now available and includes a fix for a recently
discovered security issue with the new stepmgr subsystem.
SchedMD customers were informed on October 9th and provided a patch on
request; this process is documented in our security policy. [1]
A mistake in authentication handling in stepmgr could permit an attacker
to execute processes under other users' jobs. This is limited to jobs
explicitly running with --stepmgr, or on systems that have globally
enabled stepmgr through "SlurmctldParameters=enable_stepmgr" in their
configuration. CVE-2024-48936.
Downloads are available at https://www.schedmd.com/downloads.php .
Release notes follow below.
- Tim
[1] https://www.schedmd.com/security-policy/
--
Tim Wickberg
Chief Technology Officer, SchedMD LLC
Commercial Slurm Development and Support
> * Changes in Slurm 24.05.4
> ==========================
> -- Fix generic int sort functions.
> -- Fix user look up using possible unrealized uid in the dbd.
> -- Fix FreeBSD compile issue with tls/none plugin.
> -- slurmrestd - Fix regressions that allowed slurmrestd to be run as SlurmUser
> when SlurmUser was not root.
> -- mpi/pmix fix race conditions with het jobs at step start/end which could
> make srun to hang.
> -- Fix not showing some SelectTypeParameters in scontrol show config.
> -- Avoid assert when dumping removed certain fields in JSON/YAML.
> -- Improve how shards are scheduled with affinity in mind.
> -- Fix MaxJobsAccruePU not being respected when MaxJobsAccruePA is set
> in the same QOS.
> -- Prevent backfill from planning jobs that use overlapping resources for the
> same time slot if the job's time limit is less than bf_resolution.
> -- Fix memory leak when requesting typed gres and --[cpus|mem]-per-gpu.
> -- Prevent backfill from breaking out due to "system state changed" every 30
> seconds if reservations use REPLACE or REPLACE_DOWN flags.
> -- slurmrestd - Make sure that scheduler_unset parameter defaults to true even
> when the following flags are also set: show_duplicates, skip_steps,
> disable_truncate_usage_time, run_away_jobs, whole_hetjob,
> disable_whole_hetjob, disable_wait_for_result, usage_time_as_submit_time,
> show_batch_script, and or show_job_environment. Additionaly, always make
> sure show_duplicates and disable_truncate_usage_time default to true when
> the following flags are also set: scheduler_unset, scheduled_on_submit,
> scheduled_by_main, scheduled_by_backfill, and or job_started. This effects
> the following endpoints:
> 'GET /slurmdb/v0.0.40/jobs'
> 'GET /slurmdb/v0.0.41/jobs'
> -- Ignore --json and --yaml options for scontrol show config to prevent mixing
> output types.
> -- Fix not considering nodes in reservations with Maintenance or Overlap flags
> when creating new reservations with nodecnt or when they replace down nodes.
> -- Fix suspending/resuming steps running under a 23.02 slurmstepd process.
> -- Fix options like sprio --me and squeue --me for users with a uid greater
> than 2147483647.
> -- fatal() if BlockSizes=0. This value is invalid and would otherwise cause the
> slurmctld to crash.
> -- sacctmgr - Fix issue where clearing out a preemption list using
> preempt='' would cause the given qos to no longer be preempt-able until set
> again.
> -- Fix stepmgr creating job steps concurrently.
> -- data_parser/v0.0.40 - Avoid dumping "Infinity" for NO_VAL tagged "number"
> fields.
> -- data_parser/v0.0.41 - Avoid dumping "Infinity" for NO_VAL tagged "number"
> fields.
> -- slurmctld - Fix a potential leak while updating a reservation.
> -- slurmctld - Fix state save with reservation flags when a update fails.
> -- Fix reservation update issues with parameters Accounts and Users, when
> using +/- signs.
> -- slurmrestd - Don't dump warning on empty wckeys in:
> 'GET /slurmdb/v0.0.40/config'
> 'GET /slurmdb/v0.0.41/config'
> -- Fix slurmd possibly leaving zombie processes on start up in configless when
> the initial attempt to fetch the config fails.
> -- Fix crash when trying to drain a non-existing node (possibly deleted
> before).
> -- slurmctld - fix segfault when calculating limit decay for jobs with an
> invalid association.
> -- Fix IPMI energy gathering with multiple sensors.
> -- data_parser/v0.0.39 - Remove xassert requiring errors and warnings to have a
> source string.
> -- slurmrestd - Prevent potential segfault when there is an error parsing an
> array field which could lead to a double xfree. This applies to several
> endpoints in data_parser v0.0.39, v0.0.40 and v0.0.41.
> -- scancel - Fix a regression from 23.11.6 where using both the --ctld and
> --sibling options would cancel the federated job on all clusters instead of
> only the cluster(s) specified by --sibling.
> -- accounting_storage/mysql - Fix bug when removing an association
> specified with an empty partition.
> -- Fix setting multiple partition state restore on a job correctly.
> -- Fix difference in behavior when swapping partition order in job submission.
> -- Fix security issue in stepmgr that could permit an attacker to execute
> processes under other users' jobs. CVE-2024-48936.
Available presentations from this year's SLUG event are now online.
They can be found at https://www.schedmd.com/publications/
We thank all those who presented and attended for a great event!
--
Victoria Hobson
SchedMD LLC
Vice President of Marketing
We are pleased to announce the availability of Slurm versions 24.05.3
and 23.11.10.
Version 24.05.3 fixes a potential database problem when deleting a qos.
This bug only existed in 24.05.
Both versions have fixes for jobs potentially being stuck when using
cloud nodes when some nodes are powered down, a regression in 23.11.9
and 24.05.2 that caused sattach to crash, and some other minor issues.
Slurm can be downloaded from https://www.schedmd.com/downloads.php .
--
Marshall Garey
Release Management, Support, and Development
SchedMD LLC - Commercial Slurm Development and Support
> * Changes in Slurm 24.05.3
> ==========================
> -- data_parser/v0.0.40 - Added field descriptions
> -- slurmrestd - Avoid creating new slurmdbd connection per request to
> '* /slurm/slurmctld/*/*' endpoints.
> -- Fix compilation issue with switch/hpe_slingshot plugin.
> -- Fix gres per task allocation with threads-per-core.
> -- data_parser/v0.0.41 - Added field descriptions
> -- slurmrestd - Change back generated OpenAPI schema for
> `DELETE /slurm/v0.0.40/jobs/` to RequestBody instead of using parameters
> for request. slurmrestd will continue accept endpoint requests via
> RequestBody or HTTP query.
> -- topology/tree - Fix issues with switch distance optimization.
> -- Fix potential segfault of secondary slurmctld when falling back to the
> primary when running with a JobComp plugin.
> -- Enable --json/--yaml=v0.0.39 options on client commands to dump data using
> data_parser/v0.0.39 instead or outputting nothing.
> -- switch/hpe_slingshot - Fix issue that could result in a 0 length state file.
> -- Fix unnecessary message protocol downgrade for unregistered nodes.
> -- Fix unnecessarily packing alias addrs when terminating jobs with a mix of
> non-cloud/dynamic nodes and powered down cloud/dynamic nodes.
> -- accounting_storage/mysql - Fix issue when deleting a qos that could remove
> too many commas from the qos and/or delta_qos fields of the assoc table.
> -- slurmctld - Fix memory leak when using RestrictedCoresPerGPU.
> -- Fix allowing access to reservations without MaxStartDelay set.
> -- Fix regression introduced in 24.05.0rc1 breaking srun --send-libs parsing.
> -- Fix slurmd vsize memory leak when using job submission/allocation commands
> that implicitly or explicitly use --get-user-env.
> -- slurmd - Fix node going into invalid state when using CPUSpecList and
> setting CPUs to the # of cores on a multithreaded node
> -- Fix reboot asap nodes being considered in backfill after a restart.
> -- Fix --clusters/-M queries for clusters outside of a federation when
> fed_display is configured.
> -- Fix scontrol allowing updating job with bad cpus-per-task value.
> -- sattach - Fix regression from 24.05.2 security fix leading to crash.
> -- mpi/pmix - Fix assertion when built under --enable-debug.
> * Changes in Slurm 23.11.10
> ===========================
> -- switch/hpe_slingshot - Fix issue that could result in a 0 length state file.
> -- Fix unnecessary message protocol downgrade for unregistered nodes.
> -- Fix unnecessarily packing alias addrs when terminating jobs with a mix of
> non-cloud/dynamic nodes and powered down cloud/dynamic nodes.
> -- Fix allowing access to reservations without MaxStartDelay set.
> -- Fix scontrol allowing updating job with bad cpus-per-task value.
> -- sattach - Fix regression from 23.11.9 security fix leading to crash.
Slurm versions 24.05.2, 23.11.9, and 23.02.8 are now available and
include a fix for a recently discovered security issue with the switch
plugins.
SchedMD customers were informed on July 17th and provided a patch on
request; this process is documented in our security policy. [1]
For the switch/hpe_slingshot and switch/nvidia_imex plugins, a user
could override the isolation between Slingshot VNIs or IMEX channels.
If you do not have one of these switch plugins configured, then you are
not impacted by this issue.
It is unclear what, if any, information could be accessed with access to
an unauthorized channel. This disclosure is being made out of an
abundance of caution.
If you do have one of these plugins enabled, the slurmctld must be
restarted before the slurmd daemons to avoid disruption.
Downloads are available at https://www.schedmd.com/downloads.php .
Release notes follow below.
- Tim
[1] https://www.schedmd.com/security-policy/
--
Tim Wickberg
Chief Technology Officer, SchedMD LLC
Commercial Slurm Development and Support
> * Changes in Slurm 24.05.2
> ==========================
> -- Fix energy gathering rpc counter underflow in _rpc_acct_gather_energy when
> more than 10 threads try to get energy at the same time. This prevented
> the possibility to get energy from slurmd by any step until slurmd was
> restarted, so losing energy accounting metrics in the node.
> -- accounting_storage/mysql - Fix issue where new user with wckey did not
> have a default wckey sent to the slurmctld.
> -- slurmrestd - Prevent slurmrestd segfault when handling the following
> endpoints when none of the optional parameters are specified:
> 'DELETE /slurm/v0.0.40/jobs'
> 'DELETE /slurm/v0.0.41/jobs'
> 'GET /slurm/v0.0.40/shares'
> 'GET /slurm/v0.0.41/shares'
> 'GET /slurmdb/v0.0.40/instance'
> 'GET /slurmdb/v0.0.41/instance'
> 'GET /slurmdb/v0.0.40/instances'
> 'GET /slurmdb/v0.0.41/instances'
> 'POST /slurm/v0.0.40/job/{job_id}'
> 'POST /slurm/v0.0.41/job/{job_id}'
> -- Fix IPMI energy gathering when no IPMIPowerSensors are specified in
> acct_gather.conf. This situation resulted in an accounted energy of 0
> for job steps.
> -- Fix a minor memory leak in slurmctld when updating a job dependency.
> -- scontrol,squeue - Fix regression that caused incorrect values for
> multisocket nodes at '.jobs[].job_resources.nodes.allocation' for
> 'scontrol show jobs --(json|yaml)' and 'squeue --(json|yaml)'.
> -- slurmrestd - Fix regression that caused incorrect values for
> multisocket nodes at '.jobs[].job_resources.nodes.allocation' to be dumped
> with endpoints:
> 'GET /slurm/v0.0.41/job/{job_id}'
> 'GET /slurm/v0.0.41/jobs'
> -- jobcomp/filetxt - Fix truncation of job record lines > 1024 characters.
> -- Fixed regression that prevented compilation on FreeBSD hosts.
> -- switch/hpe_slingshot - Drain node on failure to delete CXI services.
> -- Fix a performance regression from 23.11.0 in cpu frequency handling when no
> CpuFreqDef is defined.
> -- Fix one-task-per-sharing not working across multiple nodes.
> -- Fix inconsistent number of cpus when creating a reservation using the
> TRESPerNode option.
> -- data_parser/v0.0.40+ - Fix job state parsing which could break filtering.
> -- Prevent cpus-per-task to be modified in jobs where a -c value has been
> explicitly specified and the requested memory constraints implicitly
> increase the number of CPUs to allocate.
> -- slurmrestd - Fix regression where args '-s v0.0.39,dbv0.0.39' and
> '-d v0.0.39' would result in 'GET /openapi/v3' not registering as a valid
> possible query resulting in 404 errors.
> -- slurmrestd - Fix memory leak for dbv0.0.39 jobs query which occurred if the
> query parameters specified account, association, cluster, constraints,
> format, groups, job_name, partition, qos, reason, reservation, state, users,
> or wckey. This affects the following endpoints:
> 'GET /slurmdb/v0.0.39/jobs'
> -- slurmrestd - In the case the slurmdbd does not respond to a persistent
> connection init message, prevent the closed fd from being used, and instead
> emit an error or warning depending on if the connection was required.
> -- Fix 24.05.0 regression that caused the slurmdbd not to send back an error
> message if there is an error initializing a persistent connection.
> -- Reduce latency of forwarded x11 packets.
> -- Add "curr_dependency" (representing the current dependency of the job)
> and "orig_dependency" (representing the original requested dependency of
> the job) fields to the job record in job_submit.lua (for job update) and
> jobcomp.lua.
> -- Fix potential segfault of slurmctld configured with
> SlurmctldParameters=enable_rpc_queue from happening on reconfigure.
> -- Fix potential segfault of slurmctld on its shutdown when rate limitting
> is enabled.
> -- slurmrestd - Fix missing job environment for SLURM_JOB_NAME,
> SLURM_OPEN_MODE, SLURM_JOB_DEPENDENCY, SLURM_PROFILE, SLURM_ACCTG_FREQ,
> SLURM_NETWORK and SLURM_CPU_FREQ_REQ to match sbatch.
> -- Add missing bash-completions dependency to slurm-smd-client debian package.
> -- Fix bash-completions installation in debian pacakges.
> -- Fix GRES environment variable indices being incorrect when only using a
> subset of all GPUs on a node and the --gres-flags=allow-task-sharing option
> -- Add missing mariadb/mysql client package dependency to debian package.
> -- Fail the debian package build early if mysql cannot be found.
> -- Prevent scontrol from segfaulting when requesting scontrol show reservation
> --json or --yaml if there is an error retrieving reservations from the
> slurmctld.
> -- switch/hpe_slingshot - Fix security issue around managing VNI access.
> -- switch/nvidia_imex - Fix security issue managing IMEX channel access.
> -- switch/nvidia_imex - Allow for compatibility with job_container/tmpfs.
> * Changes in Slurm 23.11.9
> ==========================
> -- Fix many commands possibly reporting an "Unexpected Message Received" when
> in reality the connection timed out.
> -- Fix heterogeneous job components not being signaled with scancel --ctld and
> 'DELETE slurm/v0.0.40/jobs' if the job ids are not explicitly given,
> the heterogeneous job components match the given filters, and the
> heterogeneous job leader does not match the given filters.
> -- Fix regression from 23.02 impeding job licenses from being cleared.
> -- Move error to log_flag which made _get_joules_task error to be logged to the
> user when too many rpcs were queued in slurmd for gathering energy.
> -- slurmrestd - Prevent a slurmrestd segfault when modifying an association
> without specifying max TRES limits in the request if those TRES
> limits are currently defined in the association. This affects the following
> fields of endpoint 'POST /slurmdb/v0.0.38/associations/':
> 'associations/max/tres/per/job'
> 'associations/max/tres/per/node'
> 'associations/max/tres/total'
> 'associations/max/tres/minutes/per/job'
> 'associations/max/tres/minutes/total'
> -- Fix power_save operation after recovering from a failed reconfigure.
> -- scrun - Delay shutdown until after start requested. This caused scrun
> to never start or shutdown and hung forever when using --tty.
> -- Fix backup slurmctld potentially not running the agent when taking over as
> the primary controller.
> -- Fix primary controller not running the agent when a reconfigure of the
> slurmctld fails.
> -- jobcomp/{elasticsearch,kafka} - Avoid sending fields with invalid date/time.
> -- Fix energy gathering rpc counter underflow in _rpc_acct_gather_energy when
> more than 10 threads try to get energy at the same time. This prevented
> the possibility to get energy from slurmd by any step until slurmd was
> restarted, so losing energy accounting metrics in the node.
> -- slurmrestd - Fix memory leak for dbv0.0.39 jobs query which occurred if the
> query parameters specified account, association, cluster, constraints,
> format, groups, job_name, partition, qos, reason, reservation, state, users,
> or wckey. This affects the following endpoints:
> 'GET /slurmdb/v0.0.39/jobs'
> -- switch/hpe_slingshot - Fix security issue around managing VNI access.
> * Changes in Slurm 23.02.8
> ==========================
> -- Fix rare deadlock when a dynamic node registers at the same time that a
> once per minute background task occurs.
> -- Fix assertion in developer mode on a failed message unpack.
> -- switch/hpe_slingshot - Fix security issue around managing VNI access.
Slurm User Group (SLUG) 2024 is set for September 12-13 at the
University of Oslo in Oslo, Norway.
Registration information, abstracts, and travel recommendations can be
found here:https://slug24.splashthat.com/
The last day to register with standard pricing ($900) is this Friday,
August 2nd. After this, final registration will run until August 30th
at a price of $1100.
SLUG is the best way to interact with the Slurm community and to
interact with the SchedMD Support & Training staff.
Don't forget to register. We can't wait to see you in Oslo!
--
Victoria Hobson
SchedMD LLC
Vice President of Marketing
We are pleased to announce the availability of Slurm version 24.05.1.
This release addresses a number of minor-to-moderate issues since the
24.05 release was first announced a month ago.
Slurm can be downloaded from https://www.schedmd.com/downloads.php .
- Tim
> * Changes in Slurm 24.05.1
> ==========================
> -- Fix slurmctld and slurmdbd potentially stopping instead of performing a
> logrotate when recieving SIGUSR2 when using auth/slurm.
> -- switch/hpe_slingshot - Fix slurmctld crash when upgrading from 23.02.
> -- Fix "Could not find group" errors from validate_group() when using
> AllowGroups with large /etc/group files.
> -- Prevent an assertion in debugging builds when triggering log rotation
> in a backup slurmctld.
> -- Add AccountingStoreFlags=no_stdio which allows to not record the stdio
> paths of the job when set.
> -- slurmrestd - Prevent a slurmrestd segfault when parsing the crontab field,
> which was never usable. Now it explicitly ignores the value and emits a
> warning if it is used for the following endpoints:
> 'POST /slurm/v0.0.39/job/{job_id}'
> 'POST /slurm/v0.0.39/job/submit'
> 'POST /slurm/v0.0.40/job/{job_id}'
> 'POST /slurm/v0.0.40/job/submit'
> 'POST /slurm/v0.0.41/job/{job_id}'
> 'POST /slurm/v0.0.41/job/submit'
> 'POST /slurm/v0.0.41/job/allocate'
> -- mpi/pmi2 - Fix communication issue leading to task launch failure with
> "invalid kvs seq from node".
> -- Fix getting user environment when using sbatch with "--get-user-env" or
> "--export=" when there is a user profile script that reads /proc.
> -- Prevent slurmd from crashing if acct_gather_energy/gpu is configured but
> GresTypes is not configured.
> -- Do not log the following errors when AcctGatherEnergyType plugins are used
> but a node does not have or cannot find sensors:
> "error: _get_joules_task: can't get info from slurmd"
> "error: slurm_get_node_energy: Zero Bytes were transmitted or received"
> However, the following error will continue to be logged:
> "error: Can't get energy data. No power sensors are available. Try later"
> -- sbatch, srun - Set SLURM_NETWORK environment variable if --network is set.
> -- Fix cloud nodes not being able to forward to nodes that restarted with new
> IP addresses.
> -- Fix cwd not being set correctly when running a SPANK plugin with a
> spank_user_init() hook and the new "contain_spank" option set.
> -- slurmctld - Avoid deadlock during shutdown when auth/slurm is active.
> -- Fix segfault in slurmctld with topology/block.
> -- sacct - Fix printing of job group for job steps.
> -- scrun - Log when an invalid environment variable causes the job submission
> to be rejected.
> -- accounting_storage/mysql - Fix problem where listing or modifying an
> association when specifying a qos list could hang or take a very long time.
> -- gpu/nvml - Fix gpuutil/gpumem only tracking last GPU in step. Now,
> gpuutil/gpumem will record sums of all GPUS in the step.
> -- Fix error in scrontab jobs when using slurm.conf:PropagatePrioProcess=1.
> -- Fix slurmctld crash on a batch job submission with "--nodes 0,...".
> -- Fix dynamic IP address fanout forwarding when using auth/slurm.
> -- Restrict listening sockets in the mpi/pmix plugin and sattach to the
> SrunPortRange.
> -- slurmrestd - Limit mime types returned from query to 'GET /openapi/v3' to
> only return one mime type per serializer plugin to fix issues with OpenAPI
> client generators that are unable to handle multiple mime type aliases.
> -- Fix many commands possibly reporting an "Unexpected Message Received" when
> in reality the connection timed out.
> -- Prevent slurmctld from starting if there is not a json serializer present
> and the extra_constraints feature is enabled.
> -- Fix heterogeneous job components not being signaled with scancel --ctld and
> 'DELETE slurm/v0.0.40/jobs' if the job ids are not explicitly given,
> the heterogeneous job components match the given filters, and the
> heterogeneous job leader does not match the given filters.
> -- Fix regression from 23.02 impeding job licenses from being cleared.
> -- Move error to log_flag which made _get_joules_task error to be logged to the
> user when too many rpcs were queued in slurmd for gathering energy.
> -- For scancel --ctld and the associated rest api endpoints:
> 'DELETE /slurm/v0.0.40/jobs'
> 'DELETE /slurm/v0.0.41/jobs'
> Fix canceling the final array task in a job array when the task is pending
> and all array tasks have been split into separate job records. Previously
> this task was not canceled.
> -- Fix power_save operation after recovering from a failed reconfigure.
> -- slurmctld - Skip removing the pidfile when running under systemd. In that
> situation it is never created in the first place.
> -- Fix issue where altering the flags on a Slurm account (UsersAreCoords)
> several limits on the account's association would be set to 0 in
> Slurm's internal cache.
> -- Fix memory leak in the controller when relaying stepmgr step accounting to
> the dbd.
> -- Fix segfault when submitting stepmgr jobs within an existing allocation.
> -- Added "disable_slurm_hydra_bootstrap" as a possible MpiParams parameter in
> slurm.conf. Using this will disable env variable injection to allocations
> for the following variables: I_MPI_HYDRA_BOOTSTRAP,
> I_MPI_HYDRA_BOOTSTRAP_EXEC_EXTRA_ARGS, HYDRA_BOOTSTRAP,
> HYDRA_LAUNCHER_EXTRA_ARGS.
> -- scrun - Delay shutdown until after start requested. This caused scrun
> to never start or shutdown and hung forever when using --tty.
> -- Fix backup slurmctld potentially not running the agent when taking over as
> the primary controller.
> -- Fix primary controller not running the agent when a reconfigure of the
> slurmctld fails.
> -- slurmd - fix premature timeout waiting for REQUEST_LAUNCH_PROLOG with large
> array jobs causing node to drain.
> -- jobcomp/{elasticsearch,kafka} - Avoid sending fields with invalid date/time.
> -- jobcomp/elasticsearch - Fix slurmctld memory leak from curl usage
> -- acct_gather_profile/influxdb - Fix slurmstepd memory leak from curl usage
> -- Fix 24.05.0 regression not deleting job hash dirs after MinJobAge.
> -- Fix filtering arguments being ignored when using squeue --json.
> -- switch/nvidia_imex - Move setup call after spank_init() to allow namespace
> manipulation within the SPANK plugin.
> -- switch/nvidia_imex - Skip plugin operation if nvidia-caps-imex-channels
> device is not present rather than preventing slurmd from starting.
> -- switch/nvidia_imex - Skip plugin operation if job_container/tmpfs
> is configured due to incompatibility.
> -- switch/nvidia_imex - Remove any pre-existing channels when slurmd starts.
> -- rpc_queue - Add support for an optional rpc_queue.yaml configuration file.
We are pleased to announce the availability of Slurm version 23.11.8.
The 23.11.8 release fixes some potential crashes in slurmctld,
slurmrestd, and slurmd when using less common features; two issues in
auth/slurm; and a few other minor bugs.
Slurm can be downloaded from https://www.schedmd.com/downloads.php .
-Marshall
> -- Fix slurmctld crash when reconfiguring with a PrologSlurmctld is running.
> -- Fix slurmctld crash after a job has been resized.
> -- Fix slurmctld and slurmdbd potentially stopping instead of performing a
> logrotate when recieving SIGUSR2 when using auth/slurm.
> -- Fix not having a disabled value for keepalive CommunicationParameters in
> slurm.conf when these parameters are not set. This can log an error when
> setting a socket, for example during slurmdbd registration with ctld.
> -- switch/hpe_slingshot - Fix slurmctld crash when upgrading from 23.02.
> -- Fix "Could not find group" errors from validate_group() when using
> AllowGroups with large /etc/group files.
> -- slurmrestd - Prevent a slurmrestd segfault when parsing the crontab field,
> which was never usable. Now it explicitly ignores the value and emits a
> warning if it is used for the following endpoints:
> 'POST /slurm/v0.0.39/job/{job_id}'
> 'POST /slurm/v0.0.39/job/submit'
> 'POST /slurm/v0.0.40/job/{job_id}'
> 'POST /slurm/v0.0.40/job/submit'
> -- Fix getting user environment when using sbatch with "--get-user-env" or
> "--export=" when there is a user profile script that reads /proc.
> -- Prevent slurmd from crashing if acct_gather_energy/gpu is configured but
> GresTypes is not configured.
> -- Do not log the following errors when AcctGatherEnergyType plugins are used
> but a node does not have or cannot find sensors:
> "error: _get_joules_task: can't get info from slurmd"
> "error: slurm_get_node_energy: Zero Bytes were transmitted or received"
> However, the following error will continue to be logged:
> "error: Can't get energy data. No power sensors are available. Try later"
> -- Fix cloud nodes not being able to forward to nodes that restarted with new
> IP addresses.
> -- sacct - Fix printing of job group for job steps.
> -- Fix error in scrontab jobs when using slurm.conf:PropagatePrioProcess=1.
> -- Fix slurmctld crash on a batch job submission with "--nodes 0,...".
> -- Fix dynamic IP address fanout forwarding when using auth/slurm.
We are pleased to announce the availability of Slurm 24.05.0.
To highlight some new features in 24.05:
- Isolated Job Step management. Enabled on a job-by-job basis with the
--stepmgr option, or globally through SlurmctldParameters=enable_stepmgr.
- Federation - Allow for client command operation while SlurmDBD is
unavailable.
- New MaxTRESRunMinsPerAccount and MaxTRESRunMinsPerUser QOS limits.
- New USER_DELETE reservation flag.
- New Flags=rebootless option on Features for node_features/helpers
which indicates the given feature can be enabled without rebooting the node.
- Cloud power management options: New "max_powered_nodes=<limit>" option
in SlurmctldParamters, and new SuspendExcNodes=<nodes>:<count> syntax
allowing for <count> nodes out of a given node list to be excluded.
- StdIn/StdOut/StdErr now stored in SlurmDBD accounting records for
batch jobs.
- New switch/nvidia_imex plugin for IMEX channel management on NVIDIA
systems.
- New RestrictedCoresPerGPU option at the Node level, designed to ensure
GPU workloads always have access to a certain number of CPUs even when
nodes are running non-GPU workloads concurrently.
The Slurm documentation has also been updated to the 24.05 release.
(Older versions can be found in the archive, linked from the main
documentation page.)
Slurm can be downloaded from https://www.schedmd.com/downloads.php .
--
Marshall Garey
Release Management, Support, and Development
SchedMD LLC - Commercial Slurm Development and Support
Slurm User Group (SLUG) 2024 is set for September 12-13 at the
University of Oslo in Oslo, Norway.
Registration information and a high-level schedule can be found
here:https://slug24.splashthat.com/ The last day to register at the
early bird pricing is this Friday, May 31st.
Friday is also the deadline to submit a presentation abstract. We do
not intend to extend this deadline.
If you are interested in presenting your own usage, developments, site
report, tutorial, etc about Slurm, please fill out the following
form:https://forms.gle/N7bFo5EzwuTuKkBN7
Notifications of final presentations accepted will go out by Friday, June 14th.
--
Victoria Hobson
SchedMD LLC
Vice President of Marketing
We are pleased to announce the availability of Slurm version 23.11.7.
The 23.11.7 release fixes a few potential crashes in slurmctld when
using less common options on job submission, slurmrestd compatibility
with auth/slurm, and some additional minor and moderate severity bugs.
Slurm can be downloaded from https://www.schedmd.com/downloads.php .
-Marshall
> -- slurmrestd - Correct OpenAPI specification for
> 'GET /slurm/v0.0.40/jobs/state' having response as null.
> -- Allow running jobs on overlapping partitions if jobs don't specify -s.
> -- Fix segfault when requesting a shared gres along with an exclusive
> allocation.
> -- Fix regression in 23.02 where afternotok and afterok dependencies were
> rejected for federated jobs not running on the origin cluster of the
> submitting job.
> -- slurmctld - Disable job table locking while job state cache is active when
> replying to `squeue --only-job-state` or `GET /slurm/v0.0.40/jobs/state`.
> -- Fix sanity check when setting tres-per-task on the job allocation as well as
> the step.
> -- slurmrestd - Fix compatiblity with auth/slurm.
> -- Fix issue where TRESRunMins gets off correct value if using
> QOS UsageFactor != 1.
> -- slurmrestd - Require `user` and `association_condition` fields to be
> populated for requests to 'POST /slurmdb/v0.0.40/users_association'.
> -- Avoid a slurmctld crash with extra_constraints enabled when a job requests
> certain invalid --extra values.
> -- `scancel --ctld` and `DELETE /slurm/v0.0/40/jobs` - Fix support for job
> array expressions (e.g. 1_[3-5]). Also fix signaling a single pending array
> task (e.g. 1_10), which previously signaled the whole array job instead.
> -- Fix a possible slurmctld segfault when at some point we failed to create an
> external launcher step.
> -- Allow the slurmctld to open a connection to the slurmdbd if the first
> attempt fails due to a protocol error.
> -- mpi/cray_shasta - Fix launch for non-het-steps within a hetjob.
> -- sacct - Fix "gpuutil" TRES usage output being incorrect when using --units.
> -- Fix a rare deadlock on slurmctld shutdown or reconfigure.
> -- Fix issue that only left one thread on each core available when "CPUs=" is
> configured to total thread count on multi-threaded hardware and no other
> topology info ("Sockets=", "CoresPerSocket", etc.) is configured.
> -- Fix the external launcher step not being allocated a VNI when requested.
> -- jobcomp/kafka - Fix payload length when producing and sending a message.
> -- scrun - Avoid a crash if RunTimeDelete is called before the container
> finishes.
> -- Save the slurmd's cred_state while reconfiguring to prevent the loss job
> credentials.
Slurm User Group (SLUG) 2024 is set for September 12-13 at the
University of Oslo in Oslo, Norway.
Registration information and a high-level schedule can be found
here:https://slug24.splashthat.com/
The deadline to submit a presentation abstract is Friday, May 31st. We
do not intend to extend this deadline.
If you are interested in presenting your own usage, developments, site
report, tutorial, etc about Slurm, please fill out the following
form:https://forms.gle/N7bFo5EzwuTuKkBN7
Notifications of final presentations accepted will go out by Friday, June 14th.
--
Victoria Hobson
SchedMD LLC
Vice President of Marketing
We are pleased to announce the availability of Slurm release candidate
24.05.0rc1.
To highlight some new features coming in 24.05:
- (Optional) isolated Job Step management. Enabled on a job-by-job basis
with the --stepmgr option, or globally through
SlurmctldParameters=enable_stepmgr.
- Federation - Allow for client command operation while SlurmDBD is
unavailable.
- New MaxTRESRunMinsPerAccount and MaxTRESRunMinsPerUser QOS limits.
- New USER_DELETE reservation flag.
- New Flags=rebootless option on Features for node_features/helpers
which indicates the given feature can be enabled without rebooting the node.
- Cloud power management options: New "max_powered_nodes=<limit>" option
in SlurmctldParamters, and new SuspendExcNodes=<nodes>:<count> syntax
allowing for <count> nodes out of a given node list to be excluded.
- StdIn/StdOut/StdErr now stored in SlurmDBD accounting records for
batch jobs.
- New switch/nvidia_imex plugin for IMEX channel management on NVIDIA
systems.
- New RestrictedCoresPerGPU option at the Node level, designed to ensure
GPU workloads always have access to a certain number of CPUs even when
nodes are running non-GPU workloads concurrently.
This is the first release candidate of the upcoming 24.05 release
series, and represents the end of development for this release, and a
finalization of the RPC and state file formats.
If any issues are identified with this release candidate, please report
them through https://bugs.schedmd.com against the 24.05.x version and we
will address them before the first production 24.05.0 release is made.
Please note that the release candidates are not intended for production use.
A preview of the updated documentation can be found at
https://slurm.schedmd.com/archive/slurm-master/ .
Slurm can be downloaded from https://www.schedmd.com/downloads.php .
--
Marshall Garey
Release Management, Support, and Development
SchedMD LLC - Commercial Slurm Development and Support
We are pleased to announce the availability of Slurm version 23.11.6.
The 23.11.6 release includes two different problems with the
priority/multifactor plugin: a crash and a miscalculation of
AssocGrpCPURunMinutes after a slurmctld reconfiguration/restart.
The wsrep_on errors that sites running MySQL or older MariaDB should
happen much less frequently and has a clarifying statement when it
is an innocuous error.
Slurm can be downloaded from https://www.schedmd.com/downloads.php .
-Marshall
> * Changes in Slurm 23.11.6
> ==========================
> -- Avoid limiting sockets per node to one when using gres enforce-binding.
> -- slurmrestd - Avoid permission denied errors when attempting to listen on
> the same port multiple times.
> -- Fix GRES reservations where the GRES has no topology
> (no cores= in gres.conf).
> -- Ensure that thread_id_rpc is gone before priority_g_fini().
> -- Fix scontrol reboot timeout removing drain state from nodes.
> -- squeue - Print header on empty reponse to `--only-job-state`.
> -- Fix slurmrestd not ending job properly when xauth is not present and a x11
> job is sent.
> -- Add experimental job state caching with
> SchedulerParameters=enable_job_state_cache to speed up querying job states
> with squeue --only-job-state.
> -- slurmrestd - Correct dumping of invalid ArrayJobIds returned from
> 'GET /slurm/v0.0.40/jobs/state'.
> -- squeue - Correct dumping of invalid ArrayJobIds returned from
> `squeue --only-job-state --{json|yaml}`.
> -- If scancel --ctld is not used with --interactive, --sibling, or specific
> step ids, then this option issues a single request to the slurmctld to
> signal all jobs matching the specified filters. This greatly improves
> the performance of slurmctld and scancel. The updated --ctld option also
> fixes issues with the --partition or --reservation scancel options for jobs
> that requested multiple partitions or reservations.
> -- slurmrestd - Give EINVAL error when failing to parse signal name to numeric
> signal.
> -- slurmrestd - Allow ContentBody for all methods per RFC7230 even if ignored.
> -- slurmrestd - Add 'DELETE /slurm/v0.0.40/jobs' endpoint to allow bulk job
> signaling via slurmctld.
> -- Fix combination of --nodelist and --exclude not always respecting the
> excluded node list.
> -- Fix jobs incorrectly allocating nodes exclusively when started on a
> partition that doesn't enforce it. This could happen if a multi-partition
> job doesn't specify --exclusive and is evaluated first on a partition
> configured with OverSubscribe=EXCLUSIVE but ends up starting in a partition
> configured with OverSubscribe!=EXCLUSIVE evaluated afterwards.
> -- Setting GLOB_SILENCE flag no longer exposes old bugged behavior.
> -- Fix associations AssocGrpCPURunMinutes being incorrectly computed for
> running jobs after a controller reconfiguration/restart.
> -- Fix scheduling jobs that request --gpus and nodes have different node
> weights and different numbers of gpus.
> -- slurmrestd - Add "NO_CRON_JOBS" as possible flag value to the following:
> 'DELETE /slurm/v0.0.40/jobs' flags field.
> 'DELETE /slurm/v0.0.40/job/{job_id}?flags=' flags query parameter.
> -- Fix scontrol segfault/assert failure if the TRESPerNode parameter is used
> when creating reservations.
> -- Avoid checking for wsrep_on when restoring streaming replication settings.
> -- Clarify in the logs that error "1193 Unknown system variable 'wsrep_on'" is
> innocuous.
> -- accounting_storage/mysql - Fix problem when loading reservations from an
> archive dump.
> -- slurmdbd - Fix minor race condition when sending updates to a shutdown
> slurmctld.
> -- slurmctld - Fix invalid refusal of a reservation update.
> -- openapi - Fix memory leak of /meta/slurm/cluster response field.
> -- Fix memory leak when using auth/slurm and AuthInfo=use_client_ids.
--
Marshall Garey
Release Management, Support, and Development
SchedMD LLC - Commercial Slurm Development and Support
Slurm User Group (SLUG) 2024 is set for September 12-13 at the
University of Oslo in Oslo, Norway.
Registration information and a high-level schedule can be found here:
https://slug24.splashthat.com/
We invite all interested attendees to submit a presentation abstract
to be given at SLUG. Presentation content can be in the form of a
tutorial, technical presentation or site report.
SLUG 2024 is sponsored and organized by the University of Oslo and
SchedMD. This international event is open to those who want to:
Learn more about Slurm, a highly scalable resource manager and job scheduler
- Share their knowledge and experience with other users and administrators
- Get detailed information about the latest features and developments
- Share requirements and discuss future developments
Everyone who wants to present their own usage, developments, site
report, or tutorial about Slurm is invited to submit abstract details
here: https://forms.gle/N7bFo5EzwuTuKkBN7
Abstracts are due Friday, May 31st and notifications of acceptance
will go out by Friday, June 14th.
--
Victoria Hobson
SchedMD LLC
Vice President of Marketing