[slurm-users] cpu-bind=MASK at output files
Gestió Servidors
sysadmin.caos at uab.cat
Tue Jun 27 11:45:22 UTC 2023
Hello,
Running this simple script:
#!/bin/bash
#
#SBATCH --job-name=mega_job
#SBATCH --output=mega_job.out
#SBATCH --tasks=3
#SBATCH --array=0-5
#SBATCH --partition=cuda.q
echo "STARTING"
srun echo "hello world" >> file_${SLURM_ARRAY_TASK_ID}.out
echo "ENDING"
I always get this output:
STARTING
STARTING
cpu-bind=MASK - aoclsd, task 0 0 [14072]: mask 0xc00c00 set
STARTING
cpu-bind=MASK - aoclsd, task 0 0 [14080]: mask 0xc00c set
cpu-bind=MASK - aoclsd, task 0 0 [14081]: mask 0x30030 set
cpu-bind=MASK - aoclsd, task 1 1 [14136]: mask 0xc00c0 set
STARTING
cpu-bind=MASK - aoclsd, task 0 0 [14144]: mask 0x3003 set
cpu-bind=MASK - aoclsd, task 1 1 [14145]: mask 0x3003 set
cpu-bind=MASK - aoclsd, task 2 2 [14150]: mask 0x3003 set
cpu-bind=MASK - aoclsd, task 2 2 [14158]: mask 0x300300 set
cpu-bind=MASK - aoclsd, task 2 2 [14137]: mask 0xc00c0 set
cpu-bind=MASK - aoclsd, task 0 0 [14135]: mask 0xc00c0 set
STARTING
STARTING
cpu-bind=MASK - aoclsd, task 0 0 [14156]: mask 0x300300 set
cpu-bind=MASK - aoclsd, task 1 1 [14157]: mask 0x300300 set
cpu-bind=MASK - aoclsd, task 2 2 [14175]: mask 0xc00c00 set
cpu-bind=MASK - aoclsd, task 1 1 [14174]: mask 0xc00c00 set
cpu-bind=MASK - aoclsd, task 0 0 [14173]: mask 0xc00c00 set
cpu-bind=MASK - aoclsd, task 1 1 [14197]: mask 0xc00c set
cpu-bind=MASK - aoclsd, task 0 0 [14196]: mask 0xc00c set
cpu-bind=MASK - aoclsd, task 2 2 [14198]: mask 0xc00c set
cpu-bind=MASK - aoclsd, task 0 0 [14206]: mask 0x30030 set
cpu-bind=MASK - aoclsd, task 1 1 [14207]: mask 0x30030 set
cpu-bind=MASK - aoclsd, task 2 2 [14208]: mask 0x30030 set
ENDING
ENDING
ENDING
ENDING
ENDING
ENDING
As you can see, it appears some "cpu-bind=MASK" that I would like to prevent them from appearing in the output file. I have reviewed configuration files and now I'm going to copy here main lines from my slurmd.conf file:
ControlMachine=my_server
ProctrackType=proctrack/linuxproc
AuthType=auth/munge
SwitchType=switch/none
TaskPlugin=task/none,task/affinity,task/cgroup
TaskPluginParam=none
DebugFlags=NO_CONF_HASH,BackfillMap,SelectType,Steps,TraceJobs
PropagateResourceLimits=ALL
FastSchedule=1
SchedulerType=sched/backfill
SelectType=select/cons_res
SelectTypeParameters=CR_Core
AccountingStorageType=accounting_storage/slurmdbd
JobCompType=jobcomp/filetxt
JobCompLoc=/var/log/slurm/job_completions
JobAcctGatherType=jobacct_gather/linux
SlurmctldDebug=info
SlurmctldLogFile=/var/log/slurmctld.log
SlurmdDebug=info
SlurmdLogFile=/var/log/slurmd.log
AccountingStorageEnforce=associations,limits,qos
AccountingStorageHost=my_server
AccountingStorageLoc=/var/log/slurm/slurm_job_accounting.txt
GresTypes=gpu
KillOnBadExit=1
OverTimeLimit=2
TCPTimeout=5
I would appreciate any help in the future.
Thanks.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.schedmd.com/pipermail/slurm-users/attachments/20230627/37d9f6ac/attachment-0001.htm>
More information about the slurm-users
mailing list