[slurm-users] cpu limit issue
John Hearns
hearnsj at googlemail.com
Wed Jul 11 02:30:14 MDT 2018
Mahmood,
I am sure you have checked this. Try running ps -eaf --forest while
a job is running.
I often find the --forest option helps to understand how batch jobs are
being run.
On 11 July 2018 at 09:12, Mahmood Naderan <mahmood.nt at gmail.com> wrote:
> >Check the Gaussian log file for mention of its using just 8 CPUs-- just
> because there are 12 CPUs available doesn't mean the program uses all of
> >them. It will scale-back if 12 isn't a good match to the problem as I
> recall.
>
>
>
> Well, in the log file, it says
>
> ******************************************
> %nprocshared=12
> Will use up to 12 processors via shared memory.
> %mem=18GB
> %chk=trimer.chk
>
> Maybe, it scales down to a good match. But I haven't seen that before.
> That was why I asked the question.
>
>
>
>
>
> One more question. Does it matter if the user specify (or not specify)
> --account in the sbatch script?
>
> [root at rocks7 ~]# sacctmgr list association format=partition,account,user,
> grptres,maxwall
> Partition Account User GrpTRES MaxWall
> ---------- ---------- ---------- ------------- -----------
> emerald z3 noor cpu=12,mem=1+ 30-00:00:00
>
>
>
> [noor at rocks7 ~]$ grep nprocshared trimer.gjf
> %nprocshared=12
> [noor at rocks7 ~]$ cat trimer.sh
> #!/bin/bash
> #SBATCH --output=trimer.out
> #SBATCH --job-name=trimer
> #SBATCH --ntasks=12
> #SBATCH --mem=18GB
> #SBATCH --partition=EMERALD
> g09 trimer.gjf
>
>
>
>
>
>
>
> Regards,
> Mahmood
>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20180711/4c6f5629/attachment.html>
More information about the slurm-users
mailing list