<div dir="ltr"><div>Hi,</div><div><br></div><div>We have a short partition to give a
reasonable waiting time for shorter jobs. We use the
job_submit/all_partitions plugin so if a user doesn't specify a
partition, it will add all the partitions.</div><div><br></div><div>The
downside of the plugin is that if a job is too long for the short
partition (or the job can't run on a partition for some other
reasons), the user will get for example a "PartitionTimeLimit" or "AccountNotAllowed" reasons instead of
"Priority" (though the job will still run eventually). If that's an issue, writing the above lua plugin might be the way to go.<br></div><div><br></div><div>Regards,<br></div><div> Yair.</div><br></div><div class="gmail_extra"><br><div class="gmail_quote">On Mon, Aug 13, 2018 at 4:46 PM, Shenglong Wang <span dir="ltr"><<a href="mailto:sw77@nyu.edu" target="_blank">sw77@nyu.edu</a>></span> wrote:<br><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex"><div style="word-wrap:break-word;line-break:after-white-space">Please try to use SLURM Lua plugin, setup two partitions, one for n06-n10 and one for all nodes, inside SLURM Lua plugs, you can assign jobs to different partitions based on requested wall time.<div><br></div><div>Best,</div><div>Shenglong</div><div><div class="h5"><div><br><blockquote type="cite"><div>On Aug 13, 2018, at 9:44 AM, Cyrus Proctor <<a href="mailto:cproctor@tacc.utexas.edu" target="_blank">cproctor@tacc.utexas.edu</a>> wrote:</div><br class="m_-9068671776869115028Apple-interchange-newline"><div>
<div bgcolor="#FFFFFF" text="#000000">
Hi Jens,<br>
<br>
Check out <a class="m_-9068671776869115028moz-txt-link-freetext" href="https://slurm.schedmd.com/reservations.html" target="_blank">https://slurm.schedmd.com/<wbr>reservations.html</a> specifically
the "
Reservations Floating Through Time" section. In your case, set a
walltime of 14 days for your partition that contains n[01-10]. Then,
create a floating reservation on node n[06-10] for n + 1 day where
"n" is always evaluated as now.<br>
<br>
If you wish to allow the user more control, then specify a "Feature"
in slurm.conf for you nodes. Something like:<br>
NodeName=n[01-05] Sockets=1 CoresPerSocket=48 ThreadsPerCore=2
State=UNKNOWN Feature=long<br>
NodeName=n[06-10] Sockets=1 CoresPerSocket=48 ThreadsPerCore=2
State=UNKNOWN Feature=short<br>
<br>
The feature is an arbitrary string that the admin sets. Then a user
could specify in their submission as something like:<br>
sbatch --constraint="long|short" batch.slurm<br>
<br>
Best,<br>
Cyrus<br>
<br>
<div class="m_-9068671776869115028moz-cite-prefix">On 08/13/2018 08:28 AM, Loris Bennett
wrote:<br>
</div>
<blockquote type="cite">
<pre>Hi Jens,
Jens Dreger <a class="m_-9068671776869115028moz-txt-link-rfc2396E" href="mailto:jens.dreger@physik.fu-berlin.de" target="_blank"><jens.dreger@physik.fu-berlin.<wbr>de></a> writes:
</pre>
<blockquote type="cite">
<pre>Hi everyone!
Is it possible to transparently assign different walltime limits
to nodes without forcing users to specify partitions when submitting
jobs?
Example: let's say I have 10 nodes. Nodes n01-n05 should be available
for jobs with a walltime up to 14 days, while n06-n10 should only
be used for jobs with a walltime limit less then 1 day. Then as long
as nodes n06-n10 have free resources, jobs with walltime <1day should
be scheduled to these nodes. If n06-n10 are full, jobs with walltime
<1day should start on n01-n05. Users should not have to specify
partitions.
Would this even be possible to do with just one partition much
like nodes with different memory size using weights to fill nodes
with less memoery first?
Background of this question is that it would be helpfull to be able
to lower the walltime for a rack of nodes, e.g. when adding this rack
to an existing cluster in order to be able to easily shut down just
this rack after one day in case of instabilities. Much like adding
N nodes to a cluster without changing anything else and have only
jobs with walltime <1day on thiese nodes in the beginning.
</pre>
</blockquote>
<pre>If you just want to reduce the allowed wall-time for a given rack, can't
you just use a maintenance reservation for the appropriate set of nodes?
Cheers,
Loris
</pre>
</blockquote>
<br>
</div>
</div></blockquote></div><br></div></div></div></blockquote></div><br></div>