[slurm-users] RFC: Slurm Tool to Automate and Track Large Job Arrays
Alex Chekholko
alex at calicolabs.com
Fri Jan 18 17:51:48 UTC 2019
Almost every place I worked built some site-specific tools for managing
jobs that some people found very useful. E.g.
https://github.com/StanfordBioinformatics/SJM
http://clusterjob.org/
There have also been some efforts to standardize this sort of thing:
https://www.commonwl.org/
I have not used any of them myself.
On Fri, Jan 18, 2019 at 9:16 AM Erik Surface <erik.surface at gmail.com> wrote:
> Hi, I am a slurm end-user needing to run ~250k jobs, each takes ~2-4 hrs.
> With the traffic on our cluster and a limit of 7000 job submissions at a
> time, it will take about a month to run the full set, if we are lucky.
>
> I built a generic tool (in bash, currently) that automates the tracking
> and submission of jobs on the system. More info here:
> https://github.com/esurface/smanage
>
> Are there other tools like this in the wild? Is this something helpful to
> this community or its end-users? Would it be worth building it out in a
> more digestible form ('C', API, etc.)?
>
> Thanks,
> Erik
>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20190118/837bb1f9/attachment.html>
More information about the slurm-users
mailing list