[slurm-users] siesta jobs with slurm, an issue

Renfro, Michael Renfro at tntech.edu
Sun Jul 22 12:43:00 MDT 2018


You’re getting the same fundamental error in both the interactive and batch version, though.

The ‘reinit: Reading from standard input’ line seemed off, since you were providing an argument for the input file. But all the references I find to running Siesta in their manual (section 3 and section 16) show something more like:

  mpirun -np 4 /share/apps/chem/siesta-4.0.2/spar/siesta < dimer1prime.fdf > dimer1prime.out

and those examples line up with the idea that Siesta reads its commands from standard input, not by literally opening an input file specified as a command-line argument.

If the version of the command using < and > works correctly, then it’s definitely not a Slurm issue, it’s an issue with how you invoked Siesta.

> On Jul 22, 2018, at 12:11 PM, Mahmood Naderan <mahmood.nt at gmail.com> wrote:
> 
> I am able to directly run the command on the node. Please note in the following output that I have pressed ^C after some minutes. So, the errors are related to ^C.
> 
> 
> [mahmood at compute-0-3 ~]$ mpirun -np 4 /share/apps/chem/siesta-4.0.2/spar/siesta dimer1prime.fdf dimer1prime.out
> Siesta Version  : v4.0.2
> Architecture    : x86_64-unknown-linux-gnu--unknown
> Compiler version: GNU Fortran (GCC) 4.8.5 20150623 (Red Hat 4.8.5-16)
> Compiler flags  : mpifort -g -O2
> PP flags        : -DMPI -DFC_HAVE_FLUSH -DFC_HAVE_ABORT
> PARALLEL version
> 
> * Running on    4 nodes in parallel
> >> Start of run:  22-JUL-2018  21:39:41
> 
>                            ***********************
>                            *  WELCOME TO SIESTA  *
>                            ***********************
> 
> reinit: Reading from standard input
> ************************** Dump of input data file ****************************
> ^C--------------------------------------------------------------------------
> MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD
> with errorcode 1.
> 
> NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
> You may or may not see output from other processes, depending on
> exactly when Open MPI kills them.
> --------------------------------------------------------------------------
> No species found!!!
> Stopping Program from Node:    0
> ************************** End of input data file *****************************
> 
> reinit: -----------------------------------------------------------------------
> reinit: System Name:
> reinit: -----------------------------------------------------------------------
> reinit: System Label: siesta
> reinit: -----------------------------------------------------------------------
> 
> initatom: Reading input for the pseudopotentials and atomic orbitals ----------
> No species found!!!
> Stopping Program from Node:    0
> 
> 
> Regards,
> Mahmood
> 
> 
> 
> On Sun, Jul 22, 2018 at 8:44 PM, Bill Barth <bbarth at tacc.utexas.edu> wrote:
> That doesn't look like a slurm problem to me necessarily. Looks like SIESTA quit of its own volition (thus the call to MPI_ABORT()). I suggest you ask your local site support to take a look or go to the SIESTA developers. I doubt you'll find any SIESTA experts here to help you. 
> 
> All I can suggest is to check that all the paths you have provided SIESTA are correct (the path to the executable is clearly fine b/c SIESTA starts, but can it fine prime.fdf?). Otherwise start with your local support team.
> 
> Best,
> Bill.
> 
> -- 
> Bill Barth, Ph.D., Director, HPC
> bbarth at tacc.utexas.edu        |   Phone: (512) 232-7069
> Office: ROC 1.435            |   Fax:   (512) 475-9445
>  



More information about the slurm-users mailing list