[test@ohpc137pbsop-sms ~]$ cat /etc/redhat-release Red Hat Enterprise Linux ComputeNode release 7.6 (Maipo) [test@ohpc137pbsop-sms ~]$ [test@ohpc137pbsop-sms ~]$ scontrol --version slurm 18.08.6 [test@ohpc137pbsop-sms ~]$ [root@ohpc137pbsop-sms ~]# grep ^NodeName /etc/slurm/slurm.conf NodeName=ohpc137pbsop-c001 Sockets=2 CoresPerSocket=4 ThreadsPerCore=1 Procs=8 State=UNKNOWN NodeName=ohpc137pbsop-c002 Sockets=2 CoresPerSocket=4 ThreadsPerCore=1 Procs=8 State=UNKNOWN NodeName=ohpc137pbsop-c003 NodeAddr=172.16.20.103 Sockets=2 CoresPerSocket=4 ThreadsPerCore=1 Procs=8 State=UNKNOWN [root@ohpc137pbsop-sms ~]# [root@ohpc137pbsop-sms ~]# grep ^Partition /etc/slurm/slurm.conf PartitionName=regnodes Nodes=ohpc137pbsop-c00[1-3] OverSubscribe=YES Default=YES State=UP [root@ohpc137pbsop-sms ~]# [root@ohpc137pbsop-sms ~]# grep ^TaskPlugin /etc/slurm/slurm.conf TaskPlugin=task/cgroup [root@ohpc137pbsop-sms ~]# [root@ohpc137pbsop-sms ~]# scontrol show config |grep SelectType SelectType = select/cons_res SelectTypeParameters = CR_CORE [root@ohpc137pbsop-sms ~]# [test@ohpc137pbsop-sms ~]$ cat /etc/slurm/cgroup.conf ### # # Slurm cgroup support configuration file # # See man slurm.conf and man cgroup.conf for further # information on cgroup configuration parameters #-- ConstrainCores=yes TaskAffinity=yes CgroupMountpoint=/cgroup CgroupAutomount=yes ConstrainRAMSpace=yes [test@ohpc137pbsop-sms ~]$ [test@ohpc137pbsop-sms ~]$ srun --nodes=1-1 --ntasks=6 --cpu-bind=v,cores cat /proc/self/status | grep Cpus_allowed_list task/cgroup: task[1] not enough Core objects (4 < 6), disabling affinity task/cgroup: task[3] not enough Core objects (4 < 6), disabling affinity task/cgroup: task[4] not enough Core objects (4 < 6), disabling affinity task/cgroup: task[0] not enough Core objects (4 < 6), disabling affinity task/cgroup: task[2] not enough Core objects (4 < 6), disabling affinity task/cgroup: task[5] not enough Core objects (4 < 6), disabling affinity Cpus_allowed_list: 0-2,5,24,26 Cpus_allowed_list: 0-2,5,24,26 Cpus_allowed_list: 0-2,5,24,26 Cpus_allowed_list: 0-2,5,24,26 Cpus_allowed_list: 0-2,5,24,26 Cpus_allowed_list: 0-2,5,24,26 [test@ohpc137pbsop-sms ~]$ [root@ohpc137pbsop-sms ~]# scontrol show job 275 JobId=275 JobName=cat UserId=test(1001) GroupId=test(1001) MCS_label=N/A Priority=4294901755 Nice=0 Account=test QOS=normal JobState=COMPLETED Reason=None Dependency=(null) Requeue=1 Restarts=0 BatchFlag=0 Reboot=0 ExitCode=0:0 RunTime=00:00:00 TimeLimit=UNLIMITED TimeMin=N/A SubmitTime=2019-11-20T17:33:32 EligibleTime=2019-11-20T17:33:32 AccrueTime=Unknown StartTime=2019-11-20T17:33:32 EndTime=2019-11-20T17:33:32 Deadline=N/A PreemptTime=None SuspendTime=None SecsPreSuspend=0 LastSchedEval=2019-11-20T17:33:32 Partition=regnodes AllocNode:Sid=ohpc137pbsop-sms:111664 ReqNodeList=(null) ExcNodeList=(null) NodeList=ohpc137pbsop-c001 BatchHost=ohpc137pbsop-c001 NumNodes=1 NumCPUs=6 NumTasks=6 CPUs/Task=1 ReqB:S:C:T=0:0:*:* TRES=cpu=6,node=1,billing=6 Socks/Node=* NtasksPerN:B:S:C=0:0:*:* CoreSpec=* MinCPUsNode=1 MinMemoryNode=0 MinTmpDiskNode=0 Features=(null) DelayBoot=00:00:00 OverSubscribe=OK Contiguous=0 Licenses=(null) Network=(null) Command=cat WorkDir=/home/test Power= [root@ohpc137pbsop-sms ~]# [test@ohpc137pbsop-sms ~]$ srun --nodes=1-1 --ntasks=6 --cpu-bind=v,cores printenv|grep SLURM task/cgroup: task[3] not enough Core objects (4 < 6), disabling affinity task/cgroup: task[4] not enough Core objects (4 < 6), disabling affinity task/cgroup: task[0] not enough Core objects (4 < 6), disabling affinity task/cgroup: task[1] not enough Core objects (4 < 6), disabling affinity task/cgroup: task[2] not enough Core objects (4 < 6), disabling affinity task/cgroup: task[5] not enough Core objects (4 < 6), disabling affinity SLURM_PRIO_PROCESS=0 SLURM_UMASK=0002 SLURM_CLUSTER_NAME=linux SLURM_SUBMIT_DIR=/home/test SLURM_SUBMIT_HOST=ohpc137pbsop-sms.soft.fujitsu.com SLURM_JOB_NAME=printenv SLURM_JOB_CPUS_PER_NODE=6 SLURM_NTASKS=6 SLURM_NPROCS=6 SLURM_JOB_ID=276 SLURM_JOBID=276 SLURM_STEP_ID=0 SLURM_STEPID=0 SLURM_NNODES=1 SLURM_JOB_NUM_NODES=1 SLURM_NODELIST=ohpc137pbsop-c001 SLURM_JOB_PARTITION=regnodes SLURM_TASKS_PER_NODE=6 SLURM_SRUN_COMM_PORT=37316 SLURM_JOB_ACCOUNT=test SLURM_JOB_QOS=normal SLURM_WORKING_CLUSTER=linux:ohpc137pbsop-sms:6817:8448 SLURM_JOB_NODELIST=ohpc137pbsop-c001 SLURM_STEP_NODELIST=ohpc137pbsop-c001 SLURM_STEP_NUM_NODES=1 SLURM_STEP_NUM_TASKS=6 SLURM_STEP_TASKS_PER_NODE=6 SLURM_STEP_LAUNCHER_PORT=37316 SLURM_SRUN_COMM_HOST=172.16.20.104 SLURM_TOPOLOGY_ADDR=ohpc137pbsop-c001 SLURM_TOPOLOGY_ADDR_PATTERN=node SLURM_CPUS_ON_NODE=6 SLURM_CPU_BIND=verbose,cores SLURM_CPU_BIND_LIST= SLURM_CPU_BIND_TYPE=cores SLURM_CPU_BIND_VERBOSE=verbose SLURM_TASK_PID=130986 SLURM_NODEID=0 SLURM_PROCID=0 SLURM_LOCALID=0 SLURM_LAUNCH_NODE_IPADDR=172.16.20.104 SLURM_GTIDS=0,1,2,3,4,5 SLURM_CHECKPOINT_IMAGE_DIR=/var/slurm/checkpoint SLURM_JOB_UID=1001 SLURM_JOB_USER=test SLURM_JOB_GID=1001 SLURMD_NODENAME=ohpc137pbsop-c001 SLURM_PRIO_PROCESS=0 SLURM_UMASK=0002 SLURM_CLUSTER_NAME=linux SLURM_SUBMIT_DIR=/home/test SLURM_SUBMIT_HOST=ohpc137pbsop-sms.soft.fujitsu.com SLURM_JOB_NAME=printenv SLURM_JOB_CPUS_PER_NODE=6 SLURM_NTASKS=6 SLURM_NPROCS=6 SLURM_JOB_ID=276 SLURM_JOBID=276 SLURM_STEP_ID=0 SLURM_STEPID=0 SLURM_NNODES=1 SLURM_JOB_NUM_NODES=1 SLURM_NODELIST=ohpc137pbsop-c001 SLURM_JOB_PARTITION=regnodes SLURM_TASKS_PER_NODE=6 SLURM_SRUN_COMM_PORT=37316 SLURM_JOB_ACCOUNT=test SLURM_JOB_QOS=normal SLURM_WORKING_CLUSTER=linux:ohpc137pbsop-sms:6817:8448 SLURM_JOB_NODELIST=ohpc137pbsop-c001 SLURM_STEP_NODELIST=ohpc137pbsop-c001 SLURM_STEP_NUM_NODES=1 SLURM_STEP_NUM_TASKS=6 SLURM_STEP_TASKS_PER_NODE=6 SLURM_STEP_LAUNCHER_PORT=37316 SLURM_SRUN_COMM_HOST=172.16.20.104 SLURM_TOPOLOGY_ADDR=ohpc137pbsop-c001 SLURM_TOPOLOGY_ADDR_PATTERN=node SLURM_CPUS_ON_NODE=6 SLURM_CPU_BIND=verbose,cores SLURM_CPU_BIND_LIST= SLURM_CPU_BIND_TYPE=cores SLURM_CPU_BIND_VERBOSE=verbose SLURM_TASK_PID=130988 SLURM_NODEID=0 SLURM_PROCID=2 SLURM_LOCALID=2 SLURM_LAUNCH_NODE_IPADDR=172.16.20.104 SLURM_GTIDS=0,1,2,3,4,5 SLURM_CHECKPOINT_IMAGE_DIR=/var/slurm/checkpoint SLURM_JOB_UID=1001 SLURM_JOB_USER=test SLURM_JOB_GID=1001 SLURMD_NODENAME=ohpc137pbsop-c001 SLURM_PRIO_PROCESS=0 SLURM_UMASK=0002 SLURM_CLUSTER_NAME=linux SLURM_SUBMIT_DIR=/home/test SLURM_SUBMIT_HOST=ohpc137pbsop-sms.soft.fujitsu.com SLURM_JOB_NAME=printenv SLURM_JOB_CPUS_PER_NODE=6 SLURM_NTASKS=6 SLURM_NPROCS=6 SLURM_JOB_ID=276 SLURM_JOBID=276 SLURM_STEP_ID=0 SLURM_STEPID=0 SLURM_NNODES=1 SLURM_JOB_NUM_NODES=1 SLURM_NODELIST=ohpc137pbsop-c001 SLURM_JOB_PARTITION=regnodes SLURM_TASKS_PER_NODE=6 SLURM_SRUN_COMM_PORT=37316 SLURM_JOB_ACCOUNT=test SLURM_JOB_QOS=normal SLURM_WORKING_CLUSTER=linux:ohpc137pbsop-sms:6817:8448 SLURM_JOB_NODELIST=ohpc137pbsop-c001 SLURM_STEP_NODELIST=ohpc137pbsop-c001 SLURM_STEP_NUM_NODES=1 SLURM_STEP_NUM_TASKS=6 SLURM_STEP_TASKS_PER_NODE=6 SLURM_STEP_LAUNCHER_PORT=37316 SLURM_SRUN_COMM_HOST=172.16.20.104 SLURM_TOPOLOGY_ADDR=ohpc137pbsop-c001 SLURM_TOPOLOGY_ADDR_PATTERN=node SLURM_CPUS_ON_NODE=6 SLURM_CPU_BIND=verbose,cores SLURM_CPU_BIND_LIST= SLURM_CPU_BIND_TYPE=cores SLURM_CPU_BIND_VERBOSE=verbose SLURM_TASK_PID=130990 SLURM_NODEID=0 SLURM_PROCID=4 SLURM_LOCALID=4 SLURM_LAUNCH_NODE_IPADDR=172.16.20.104 SLURM_GTIDS=0,1,2,3,4,5 SLURM_CHECKPOINT_IMAGE_DIR=/var/slurm/checkpoint SLURM_JOB_UID=1001 SLURM_JOB_USER=test SLURM_JOB_GID=1001 SLURMD_NODENAME=ohpc137pbsop-c001 SLURM_PRIO_PROCESS=0 SLURM_UMASK=0002 SLURM_CLUSTER_NAME=linux SLURM_SUBMIT_DIR=/home/test SLURM_SUBMIT_HOST=ohpc137pbsop-sms.soft.fujitsu.com SLURM_JOB_NAME=printenv SLURM_JOB_CPUS_PER_NODE=6 SLURM_NTASKS=6 SLURM_NPROCS=6 SLURM_JOB_ID=276 SLURM_JOBID=276 SLURM_STEP_ID=0 SLURM_STEPID=0 SLURM_NNODES=1 SLURM_JOB_NUM_NODES=1 SLURM_NODELIST=ohpc137pbsop-c001 SLURM_JOB_PARTITION=regnodes SLURM_TASKS_PER_NODE=6 SLURM_SRUN_COMM_PORT=37316 SLURM_JOB_ACCOUNT=test SLURM_JOB_QOS=normal SLURM_WORKING_CLUSTER=linux:ohpc137pbsop-sms:6817:8448 SLURM_JOB_NODELIST=ohpc137pbsop-c001 SLURM_STEP_NODELIST=ohpc137pbsop-c001 SLURM_STEP_NUM_NODES=1 SLURM_STEP_NUM_TASKS=6 SLURM_STEP_TASKS_PER_NODE=6 SLURM_STEP_LAUNCHER_PORT=37316 SLURM_SRUN_COMM_HOST=172.16.20.104 SLURM_TOPOLOGY_ADDR=ohpc137pbsop-c001 SLURM_TOPOLOGY_ADDR_PATTERN=node SLURM_CPUS_ON_NODE=6 SLURM_CPU_BIND=verbose,cores SLURM_CPU_BIND_LIST= SLURM_CPU_BIND_TYPE=cores SLURM_CPU_BIND_VERBOSE=verbose SLURM_TASK_PID=130987 SLURM_NODEID=0 SLURM_PROCID=1 SLURM_LOCALID=1 SLURM_LAUNCH_NODE_IPADDR=172.16.20.104 SLURM_GTIDS=0,1,2,3,4,5 SLURM_CHECKPOINT_IMAGE_DIR=/var/slurm/checkpoint SLURM_JOB_UID=1001 SLURM_JOB_USER=test SLURM_JOB_GID=1001 SLURMD_NODENAME=ohpc137pbsop-c001 SLURM_PRIO_PROCESS=0 SLURM_UMASK=0002 SLURM_CLUSTER_NAME=linux SLURM_SUBMIT_DIR=/home/test SLURM_SUBMIT_HOST=ohpc137pbsop-sms.soft.fujitsu.com SLURM_JOB_NAME=printenv SLURM_JOB_CPUS_PER_NODE=6 SLURM_NTASKS=6 SLURM_NPROCS=6 SLURM_JOB_ID=276 SLURM_JOBID=276 SLURM_STEP_ID=0 SLURM_STEPID=0 SLURM_NNODES=1 SLURM_JOB_NUM_NODES=1 SLURM_NODELIST=ohpc137pbsop-c001 SLURM_JOB_PARTITION=regnodes SLURM_TASKS_PER_NODE=6 SLURM_SRUN_COMM_PORT=37316 SLURM_JOB_ACCOUNT=test SLURM_JOB_QOS=normal SLURM_WORKING_CLUSTER=linux:ohpc137pbsop-sms:6817:8448 SLURM_JOB_NODELIST=ohpc137pbsop-c001 SLURM_STEP_NODELIST=ohpc137pbsop-c001 SLURM_STEP_NUM_NODES=1 SLURM_STEP_NUM_TASKS=6 SLURM_STEP_TASKS_PER_NODE=6 SLURM_STEP_LAUNCHER_PORT=37316 SLURM_SRUN_COMM_HOST=172.16.20.104 SLURM_TOPOLOGY_ADDR=ohpc137pbsop-c001 SLURM_TOPOLOGY_ADDR_PATTERN=node SLURM_CPUS_ON_NODE=6 SLURM_CPU_BIND=verbose,cores SLURM_CPU_BIND_LIST= SLURM_CPU_BIND_TYPE=cores SLURM_CPU_BIND_VERBOSE=verbose SLURM_TASK_PID=130989 SLURM_NODEID=0 SLURM_PROCID=3 SLURM_LOCALID=3 SLURM_LAUNCH_NODE_IPADDR=172.16.20.104 SLURM_GTIDS=0,1,2,3,4,5 SLURM_CHECKPOINT_IMAGE_DIR=/var/slurm/checkpoint SLURM_JOB_UID=1001 SLURM_JOB_USER=test SLURM_JOB_GID=1001 SLURMD_NODENAME=ohpc137pbsop-c001 SLURM_PRIO_PROCESS=0 SLURM_UMASK=0002 SLURM_CLUSTER_NAME=linux SLURM_SUBMIT_DIR=/home/test SLURM_SUBMIT_HOST=ohpc137pbsop-sms.soft.fujitsu.com SLURM_JOB_NAME=printenv SLURM_JOB_CPUS_PER_NODE=6 SLURM_NTASKS=6 SLURM_NPROCS=6 SLURM_JOB_ID=276 SLURM_JOBID=276 SLURM_STEP_ID=0 SLURM_STEPID=0 SLURM_NNODES=1 SLURM_JOB_NUM_NODES=1 SLURM_NODELIST=ohpc137pbsop-c001 SLURM_JOB_PARTITION=regnodes SLURM_TASKS_PER_NODE=6 SLURM_SRUN_COMM_PORT=37316 SLURM_JOB_ACCOUNT=test SLURM_JOB_QOS=normal SLURM_WORKING_CLUSTER=linux:ohpc137pbsop-sms:6817:8448 SLURM_JOB_NODELIST=ohpc137pbsop-c001 SLURM_STEP_NODELIST=ohpc137pbsop-c001 SLURM_STEP_NUM_NODES=1 SLURM_STEP_NUM_TASKS=6 SLURM_STEP_TASKS_PER_NODE=6 SLURM_STEP_LAUNCHER_PORT=37316 SLURM_SRUN_COMM_HOST=172.16.20.104 SLURM_TOPOLOGY_ADDR=ohpc137pbsop-c001 SLURM_TOPOLOGY_ADDR_PATTERN=node SLURM_CPUS_ON_NODE=6 SLURM_CPU_BIND=verbose,cores SLURM_CPU_BIND_LIST= SLURM_CPU_BIND_TYPE=cores SLURM_CPU_BIND_VERBOSE=verbose SLURM_TASK_PID=130991 SLURM_NODEID=0 SLURM_PROCID=5 SLURM_LOCALID=5 SLURM_LAUNCH_NODE_IPADDR=172.16.20.104 SLURM_GTIDS=0,1,2,3,4,5 SLURM_CHECKPOINT_IMAGE_DIR=/var/slurm/checkpoint SLURM_JOB_UID=1001 SLURM_JOB_USER=test SLURM_JOB_GID=1001 SLURMD_NODENAME=ohpc137pbsop-c001 [test@ohpc137pbsop-sms ~]$ [test@ohpc137pbsop-sms ~]$ srun --nodes=1-3 lscpu Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Byte Order: Little Endian CPU(s): 48 On-line CPU(s) list: 0-47 Thread(s) per core: 2 Core(s) per socket: 12 Socket(s): 2 NUMA node(s): 4 Vendor ID: GenuineIntel CPU family: 6 Model: 85 Model name: Intel(R) Xeon(R) Gold 6126 CPU @ 2.60GHz Stepping: 4 CPU MHz: 999.914 CPU max MHz: 3700.0000 CPU min MHz: 1000.0000 BogoMIPS: 5200.00 Virtualization: VT-x L1d cache: 32K L1i cache: 32K L2 cache: 1024K L3 cache: 19712K NUMA node0 CPU(s): 0-2,5-7,24-26,29-31 NUMA node1 CPU(s): 3,4,8-11,27,28,32-35 NUMA node2 CPU(s): 12-15,19,20,36-39,43,44 NUMA node3 CPU(s): 16-18,21-23,40-42,45-47 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch epb cat_l3 cdp_l3 intel_ppin intel_pt ssbd mba ibrs ibpb stibp tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm cqm mpx rdt_a avx512f avx512dq rdseed adx smap clflushopt clwb avx512cd avx512bw avx512vl xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local dtherm ida arat pln pts hwp hwp_act_window hwp_epp hwp_pkg_req pku ospke spec_ctrl intel_stibp flush_l1d Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Byte Order: Little Endian CPU(s): 48 On-line CPU(s) list: 0-47 Thread(s) per core: 2 Core(s) per socket: 12 Socket(s): 2 NUMA node(s): 4 Vendor ID: GenuineIntel CPU family: 6 Model: 85 Model name: Intel(R) Xeon(R) Gold 6126 CPU @ 2.60GHz Stepping: 4 CPU MHz: 1000.073 CPU max MHz: 3700.0000 CPU min MHz: 1000.0000 BogoMIPS: 5200.00 Virtualization: VT-x L1d cache: 32K L1i cache: 32K L2 cache: 1024K L3 cache: 19712K NUMA node0 CPU(s): 0-2,5-7,24-26,29-31 NUMA node1 CPU(s): 3,4,8-11,27,28,32-35 NUMA node2 CPU(s): 12-14,18-20,36-38,42-44 NUMA node3 CPU(s): 15-17,21-23,39-41,45-47 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch epb cat_l3 cdp_l3 intel_ppin intel_pt ssbd mba ibrs ibpb stibp tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm cqm mpx rdt_a avx512f avx512dq rdseed adx smap clflushopt clwb avx512cd avx512bw avx512vl xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local dtherm ida arat pln pts hwp hwp_act_window hwp_epp hwp_pkg_req pku ospke spec_ctrl intel_stibp flush_l1d Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Byte Order: Little Endian CPU(s): 48 On-line CPU(s) list: 0-47 Thread(s) per core: 2 Core(s) per socket: 12 Socket(s): 2 NUMA node(s): 4 Vendor ID: GenuineIntel CPU family: 6 Model: 85 Model name: Intel(R) Xeon(R) Gold 6126 CPU @ 2.60GHz Stepping: 4 CPU MHz: 1617.700 CPU max MHz: 3700.0000 CPU min MHz: 1000.0000 BogoMIPS: 5200.00 Virtualization: VT-x L1d cache: 32K L1i cache: 32K L2 cache: 1024K L3 cache: 19712K NUMA node0 CPU(s): 0-3,7,8,24-27,31,32 NUMA node1 CPU(s): 4-6,9-11,28-30,33-35 NUMA node2 CPU(s): 12-14,18-20,36-38,42-44 NUMA node3 CPU(s): 15-17,21-23,39-41,45-47 Flags: fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 monitor ds_cpl vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 3dnowprefetch epb cat_l3 cdp_l3 intel_ppin intel_pt ssbd mba ibrs ibpb stibp tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 hle avx2 smep bmi2 erms invpcid rtm cqm mpx rdt_a avx512f avx512dq rdseed adx smap clflushopt clwb avx512cd avx512bw avx512vl xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc cqm_mbm_total cqm_mbm_local dtherm ida arat pln pts hwp hwp_act_window hwp_epp hwp_pkg_req pku ospke spec_ctrl intel_stibp flush_l1d [test@ohpc137pbsop-sms ~]$ [root@ohpc137pbsop-sms ~]# scontrol show config Configuration data as of 2019-11-20T17:19:50 AccountingStorageBackupHost = (null) AccountingStorageEnforce = none AccountingStorageHost = localhost AccountingStorageLoc = N/A AccountingStoragePort = 6819 AccountingStorageTRES = cpu,mem,energy,node,billing,fs/disk,vmem,pages AccountingStorageType = accounting_storage/slurmdbd AccountingStorageUser = N/A AccountingStoreJobComment = Yes AcctGatherEnergyType = acct_gather_energy/none AcctGatherFilesystemType = acct_gather_filesystem/none AcctGatherInterconnectType = acct_gather_interconnect/none AcctGatherNodeFreq = 0 sec AcctGatherProfileType = acct_gather_profile/none AllowSpecResourcesUsage = 0 AuthInfo = (null) AuthType = auth/munge BatchStartTimeout = 10 sec BOOT_TIME = 2019-11-20T17:19:40 BurstBufferType = (null) CheckpointType = checkpoint/none ClusterName = linux CommunicationParameters = (null) CompleteWait = 0 sec CoreSpecPlugin = core_spec/none CpuFreqDef = Unknown CpuFreqGovernors = Performance,OnDemand,UserSpace CryptoType = crypto/munge DebugFlags = (null) DefMemPerNode = UNLIMITED DisableRootJobs = No EioTimeout = 60 EnforcePartLimits = NO Epilog = (null) EpilogMsgTime = 2000 usec EpilogSlurmctld = (null) ExtSensorsType = ext_sensors/none ExtSensorsFreq = 0 sec FastSchedule = 1 FederationParameters = (null) FirstJobId = 1 GetEnvTimeout = 2 sec GresTypes = (null) GroupUpdateForce = 1 GroupUpdateTime = 600 sec HASH_VAL = Match HealthCheckInterval = 0 sec HealthCheckNodeState = ANY HealthCheckProgram = (null) InactiveLimit = 0 sec JobAcctGatherFrequency = 30 JobAcctGatherType = jobacct_gather/linux JobAcctGatherParams = (null) JobCheckpointDir = /var/slurm/checkpoint JobCompHost = localhost JobCompLoc = /var/log/slurm_jobcomp.log JobCompPort = 0 JobCompType = jobcomp/none JobCompUser = root JobContainerType = job_container/none JobCredentialPrivateKey = (null) JobCredentialPublicCertificate = (null) JobDefaults = (null) JobFileAppend = 0 JobRequeue = 1 JobSubmitPlugins = (null) KeepAliveTime = SYSTEM_DEFAULT KillOnBadExit = 0 KillWait = 30 sec LaunchParameters = (null) LaunchType = launch/slurm Layouts = Licenses = (null) LicensesUsed = (null) LogTimeFormat = iso8601_ms MailDomain = (null) MailProg = /bin/mail MaxArraySize = 1001 MaxJobCount = 10000 MaxJobId = 67043328 MaxMemPerNode = UNLIMITED MaxStepCount = 40000 MaxTasksPerNode = 512 MCSPlugin = mcs/none MCSParameters = (null) MemLimitEnforce = No MessageTimeout = 10 sec MinJobAge = 300 sec MpiDefault = none MpiParams = (null) MsgAggregationParams = (null) NEXT_JOB_ID = 271 NodeFeaturesPlugins = (null) OverTimeLimit = 0 min PluginDir = /usr/lib64/slurm PlugStackConfig = /etc/slurm/plugstack.conf PowerParameters = (null) PowerPlugin = PreemptMode = OFF PreemptType = preempt/none PriorityParameters = (null) PriorityType = priority/basic PrivateData = none ProctrackType = proctrack/cgroup Prolog = (null) PrologEpilogTimeout = 65534 PrologSlurmctld = (null) PrologFlags = (null) PropagatePrioProcess = 0 PropagateResourceLimits = ALL PropagateResourceLimitsExcept = (null) RebootProgram = (null) ReconfigFlags = (null) RequeueExit = (null) RequeueExitHold = (null) ResumeFailProgram = (null) ResumeProgram = (null) ResumeRate = 300 nodes/min ResumeTimeout = 60 sec ResvEpilog = (null) ResvOverRun = 0 min ResvProlog = (null) ReturnToService = 1 RoutePlugin = route/default SallocDefaultCommand = (null) SbcastParameters = (null) SchedulerParameters = (null) SchedulerTimeSlice = 30 sec SchedulerType = sched/backfill SelectType = select/cons_res SelectTypeParameters = CR_CORE SlurmUser = slurm(202) SlurmctldAddr = (null) SlurmctldDebug = info SlurmctldHost[0] = ohpc137pbsop-sms SlurmctldLogFile = /var/log/slurmctld.log SlurmctldPort = 6817 SlurmctldSyslogDebug = unknown SlurmctldPrimaryOffProg = (null) SlurmctldPrimaryOnProg = (null) SlurmctldTimeout = 300 sec SlurmctldParameters = (null) SlurmdDebug = info SlurmdLogFile = /var/log/slurmd.log SlurmdParameters = (null) SlurmdPidFile = /var/run/slurmd.pid SlurmdPort = 6818 SlurmdSpoolDir = /var/spool/slurm/d SlurmdSyslogDebug = unknown SlurmdTimeout = 300 sec SlurmdUser = root(0) SlurmSchedLogFile = (null) SlurmSchedLogLevel = 0 SlurmctldPidFile = /var/run/slurmctld.pid SlurmctldPlugstack = (null) SLURM_CONF = /etc/slurm/slurm.conf SLURM_VERSION = 18.08.6 SrunEpilog = (null) SrunPortRange = 0-0 SrunProlog = (null) StateSaveLocation = /var/spool/slurm/ctld SuspendExcNodes = (null) SuspendExcParts = (null) SuspendProgram = (null) SuspendRate = 60 nodes/min SuspendTime = NONE SuspendTimeout = 30 sec SwitchType = switch/none TaskEpilog = (null) TaskPlugin = task/cgroup TaskPluginParam = (null type) TaskProlog = (null) TCPTimeout = 2 sec TmpFS = /tmp TopologyParam = (null) TopologyPlugin = topology/none TrackWCKey = No TreeWidth = 50 UsePam = 0 UnkillableStepProgram = (null) UnkillableStepTimeout = 60 sec VSizeFactor = 0 percent WaitTime = 0 sec X11Parameters = (null) Cgroup Support Configuration: AllowedDevicesFile = /etc/slurm/cgroup_allowed_devices_file.conf AllowedKmemSpace = (null) AllowedRAMSpace = 100.0% AllowedSwapSpace = 0.0% CgroupAutomount = yes CgroupMountpoint = /cgroup ConstrainCores = yes ConstrainDevices = no ConstrainKmemSpace = no ConstrainRAMSpace = yes ConstrainSwapSpace = no MaxKmemPercent = 100.0% MaxRAMPercent = 100.0% MaxSwapPercent = 100.0% MemLimitThreshold = 100.0% MemoryLimitEnforcement = no MemorySwappiness = (null) MinKmemSpace = 30 MB MinRAMSpace = 30 MB TaskAffinity = yes Slurmctld(primary) at ohpc137pbsop-sms is UP [root@ohpc137pbsop-sms ~]#