[slurm-users] Slurm doesn't call mpiexec or mpirun when run through a GUI app

Reuti reuti at staff.uni-marburg.de
Fri Mar 22 16:40:00 UTC 2019


> Am 22.03.2019 um 16:20 schrieb Prentice Bisbal <pbisbal at pppl.gov>:
> 
> On 3/21/19 6:56 PM, Reuti wrote:
>> Am 21.03.2019 um 23:43 schrieb Prentice Bisbal:
>> 
>>> Slurm-users,
>>> 
>>> My users here have developed a GUI application which serves as a GUI interface to various physics codes they use. From this GUI, they can submit jobs to Slurm. On Tuesday, we upgraded Slurm from 18.08.5-2 to 18.08.6-2,and a user has reported a problem when submitting Slurm jobs through this GUI app that do not occur when the same sbatch script is submitted from sbatch on the command-line.
>>> 
>>> […]
>>> When I replaced the mpirun command with an equivalent srun command, everything works as desired, so the user can get back to work and be productive.
>>> 
>>> While srun is a suitable workaround, and is arguably the correct way to run an MPI job, I'd like to understand what is going on here. Any idea what is going wrong, or additional steps I can take to get more debug information?
>> Was an alias to `mpirun` introduced? It may cover the real application and even the `which mpirun` will return the correct value, but never be executed.
>> 
>> $ type mpirun
>> $ alias mpirun
>> 
>> may tell in the jobscript.
>> 
> Unfortunately, the script is in tcsh,

Oh, I didn't notice this – correct.


> so the 'type' command doesn't work since,

Is it really running in `tcsh`? The commands look like being generic and available in various shells. Does SLURM honor the the first line of a script and/or use a default? In Bash a function would cover the `mpirun` too.

(I'm more used to GridEngine, where this can be configured in both ways how to start the scripts.)

In "tcsh" I see a defined "jobcmd" of having some effect.

-- Reuti


>  it's a bash built-in function. I did use the 'alias' command to see all the defined aliases, and mpirun and mpiexec are not aliased. Any other ideas?
> 
> Prentice
> 
> 
> 
> 




More information about the slurm-users mailing list