[slurm-users] exporting PATH inside sbatch script
Mahmood Naderan
mahmood.nt at gmail.com
Tue Jan 1 04:05:48 MST 2019
Excuse me.
That email was sent wrongly.
Please ignore it.
Regards,
Mahmood
On Tue, Jan 1, 2019 at 2:34 PM Mahmood Naderan <mahmood.nt at gmail.com> wrote:
> Hi,
> May I know why the exported path in my sbatch script doesn't work?
>
> [mahmood at rocks7 qe]$ cat slurm_script.sh
> #!/bin/bash
> #SBATCH --ntasks=4
> #SBATCH --mem=2G
> #SBATCH --partition=RUBY
> #SBATCH --account=y4
> export PATH=$PATH:/share/apps/softwares/q-e-qe-6.2.1/bin
> mpirun pw.x mos2.in output.out
> [mahmood at rocks7 qe]$ sbatch slurm_script.sh
> Submitted batch job 243
> [mahmood at rocks7 qe]$ cat slurm-243.out
> --------------------------------------------------------------------------
> mpirun was unable to find the specified executable file, and therefore
> did not launch the job. This error was first reported for process
> rank 0; it may have occurred for other processes as well.
>
> NOTE: A common cause for this error is misspelling a mpirun command
> line parameter option (remember that mpirun interprets the first
> unrecognized command line token as the executable).
>
> Node: compute-0-2
> Executable: pw.x
> --------------------------------------------------------------------------
> 4 total processes failed to start
>
>
> That folder is shared between compute nodes as you can see below
>
> [mahmood at rocks7 qe]$ rocks run host compute-0-2 "ls
> /share/apps/softwares/q-e-qe-6.2.1/bin"
> Warning: untrusted X11 forwarding setup failed: xauth key data not
> generated
> casino2upf.x
> cpmd2upf.x
> cppp.x
> cp.x
> fhi2upf.x
> fpmd2upf.x
> interpolate.x
> iotk
> iotk_print_kinds.x
> iotk.x
> ld1.x
> manycp.x
> ncpp2upf.x
> oldcp2upf.x
> read_upf_tofile.x
> rrkj2upf.x
> test.x
> upf2casino.x
> uspp2upf.x
> vdb2upf.x
> virtual.x
> wfdd.x
>
>
>
> Any help?
>
> Regards,
> Mahmood
>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20190101/403877fa/attachment.html>
More information about the slurm-users
mailing list