Hi, All

I checked 'Example 16' of CPU Management User and Administrator Guide.
However, the following message was output.

  task/cgroup: task[1] not enough Core objects (4 < 6), disabling affinity
  task/cgroup: task[3] not enough Core objects (4 < 6), disabling affinity
  task/cgroup: task[4] not enough Core objects (4 < 6), disabling affinity
  task/cgroup: task[0] not enough Core objects (4 < 6), disabling affinity
  task/cgroup: task[2] not enough Core objects (4 < 6), disabling affinity
  task/cgroup: task[5] not enough Core objects (4 < 6), disabling affinity

See attachment file for details.

What does this message mean?

Regards,
Tomo

[test@ohpc137pbsop-sms ~]$ cat /etc/redhat-release
Red Hat Enterprise Linux ComputeNode release 7.6 (Maipo)
[test@ohpc137pbsop-sms ~]$

[test@ohpc137pbsop-sms ~]$ scontrol --version
slurm 18.08.6
[test@ohpc137pbsop-sms ~]$


[root@ohpc137pbsop-sms ~]# grep ^NodeName /etc/slurm/slurm.conf
NodeName=ohpc137pbsop-c001 Sockets=2 CoresPerSocket=4 ThreadsPerCore=1 Procs=8 
State=UNKNOWN
NodeName=ohpc137pbsop-c002 Sockets=2 CoresPerSocket=4 ThreadsPerCore=1 Procs=8 
State=UNKNOWN
NodeName=ohpc137pbsop-c003 NodeAddr=172.16.20.103 Sockets=2 CoresPerSocket=4 
ThreadsPerCore=1 Procs=8 State=UNKNOWN
[root@ohpc137pbsop-sms ~]#

[root@ohpc137pbsop-sms ~]# grep ^Partition /etc/slurm/slurm.conf
PartitionName=regnodes Nodes=ohpc137pbsop-c00[1-3] OverSubscribe=YES 
Default=YES State=UP
[root@ohpc137pbsop-sms ~]#

[root@ohpc137pbsop-sms ~]# grep ^TaskPlugin /etc/slurm/slurm.conf
TaskPlugin=task/cgroup
[root@ohpc137pbsop-sms ~]#

[root@ohpc137pbsop-sms ~]# scontrol show config |grep SelectType
SelectType              = select/cons_res
SelectTypeParameters    = CR_CORE
[root@ohpc137pbsop-sms ~]#

[test@ohpc137pbsop-sms ~]$ cat /etc/slurm/cgroup.conf
###
#
# Slurm cgroup support configuration file
#
# See man slurm.conf and man cgroup.conf for further
# information on cgroup configuration parameters
#--
ConstrainCores=yes
TaskAffinity=yes
CgroupMountpoint=/cgroup
CgroupAutomount=yes
ConstrainRAMSpace=yes
[test@ohpc137pbsop-sms ~]$

[test@ohpc137pbsop-sms ~]$ srun --nodes=1-1 --ntasks=6 --cpu-bind=v,cores cat 
/proc/self/status | grep Cpus_allowed_list
task/cgroup: task[1] not enough Core objects (4 < 6), disabling affinity
task/cgroup: task[3] not enough Core objects (4 < 6), disabling affinity
task/cgroup: task[4] not enough Core objects (4 < 6), disabling affinity
task/cgroup: task[0] not enough Core objects (4 < 6), disabling affinity
task/cgroup: task[2] not enough Core objects (4 < 6), disabling affinity
task/cgroup: task[5] not enough Core objects (4 < 6), disabling affinity
Cpus_allowed_list:      0-2,5,24,26
Cpus_allowed_list:      0-2,5,24,26
Cpus_allowed_list:      0-2,5,24,26
Cpus_allowed_list:      0-2,5,24,26
Cpus_allowed_list:      0-2,5,24,26
Cpus_allowed_list:      0-2,5,24,26
[test@ohpc137pbsop-sms ~]$

[root@ohpc137pbsop-sms ~]# scontrol show job 275
JobId=275 JobName=cat
   UserId=test(1001) GroupId=test(1001) MCS_label=N/A
   Priority=4294901755 Nice=0 Account=test QOS=normal
   JobState=COMPLETED Reason=None Dependency=(null)
   Requeue=1 Restarts=0 BatchFlag=0 Reboot=0 ExitCode=0:0
   RunTime=00:00:00 TimeLimit=UNLIMITED TimeMin=N/A
   SubmitTime=2019-11-20T17:33:32 EligibleTime=2019-11-20T17:33:32
   AccrueTime=Unknown
   StartTime=2019-11-20T17:33:32 EndTime=2019-11-20T17:33:32 Deadline=N/A
   PreemptTime=None SuspendTime=None SecsPreSuspend=0
   LastSchedEval=2019-11-20T17:33:32
   Partition=regnodes AllocNode:Sid=ohpc137pbsop-sms:111664
   ReqNodeList=(null) ExcNodeList=(null)
   NodeList=ohpc137pbsop-c001
   BatchHost=ohpc137pbsop-c001
   NumNodes=1 NumCPUs=6 NumTasks=6 CPUs/Task=1 ReqB:S:C:T=0:0:*:*
   TRES=cpu=6,node=1,billing=6
   Socks/Node=* NtasksPerN:B:S:C=0:0:*:* CoreSpec=*
   MinCPUsNode=1 MinMemoryNode=0 MinTmpDiskNode=0
   Features=(null) DelayBoot=00:00:00
   OverSubscribe=OK Contiguous=0 Licenses=(null) Network=(null)
   Command=cat
   WorkDir=/home/test
   Power=

[root@ohpc137pbsop-sms ~]#



[test@ohpc137pbsop-sms ~]$ srun --nodes=1-1 --ntasks=6 --cpu-bind=v,cores 
printenv|grep SLURM
task/cgroup: task[3] not enough Core objects (4 < 6), disabling affinity
task/cgroup: task[4] not enough Core objects (4 < 6), disabling affinity
task/cgroup: task[0] not enough Core objects (4 < 6), disabling affinity
task/cgroup: task[1] not enough Core objects (4 < 6), disabling affinity
task/cgroup: task[2] not enough Core objects (4 < 6), disabling affinity
task/cgroup: task[5] not enough Core objects (4 < 6), disabling affinity
SLURM_PRIO_PROCESS=0
SLURM_UMASK=0002
SLURM_CLUSTER_NAME=linux
SLURM_SUBMIT_DIR=/home/test
SLURM_SUBMIT_HOST=ohpc137pbsop-sms.soft.fujitsu.com
SLURM_JOB_NAME=printenv
SLURM_JOB_CPUS_PER_NODE=6
SLURM_NTASKS=6
SLURM_NPROCS=6
SLURM_JOB_ID=276
SLURM_JOBID=276
SLURM_STEP_ID=0
SLURM_STEPID=0
SLURM_NNODES=1
SLURM_JOB_NUM_NODES=1
SLURM_NODELIST=ohpc137pbsop-c001
SLURM_JOB_PARTITION=regnodes
SLURM_TASKS_PER_NODE=6
SLURM_SRUN_COMM_PORT=37316
SLURM_JOB_ACCOUNT=test
SLURM_JOB_QOS=normal
SLURM_WORKING_CLUSTER=linux:ohpc137pbsop-sms:6817:8448
SLURM_JOB_NODELIST=ohpc137pbsop-c001
SLURM_STEP_NODELIST=ohpc137pbsop-c001
SLURM_STEP_NUM_NODES=1
SLURM_STEP_NUM_TASKS=6
SLURM_STEP_TASKS_PER_NODE=6
SLURM_STEP_LAUNCHER_PORT=37316
SLURM_SRUN_COMM_HOST=172.16.20.104
SLURM_TOPOLOGY_ADDR=ohpc137pbsop-c001
SLURM_TOPOLOGY_ADDR_PATTERN=node
SLURM_CPUS_ON_NODE=6
SLURM_CPU_BIND=verbose,cores
SLURM_CPU_BIND_LIST=
SLURM_CPU_BIND_TYPE=cores
SLURM_CPU_BIND_VERBOSE=verbose
SLURM_TASK_PID=130986
SLURM_NODEID=0
SLURM_PROCID=0
SLURM_LOCALID=0
SLURM_LAUNCH_NODE_IPADDR=172.16.20.104
SLURM_GTIDS=0,1,2,3,4,5
SLURM_CHECKPOINT_IMAGE_DIR=/var/slurm/checkpoint
SLURM_JOB_UID=1001
SLURM_JOB_USER=test
SLURM_JOB_GID=1001
SLURMD_NODENAME=ohpc137pbsop-c001
SLURM_PRIO_PROCESS=0
SLURM_UMASK=0002
SLURM_CLUSTER_NAME=linux
SLURM_SUBMIT_DIR=/home/test
SLURM_SUBMIT_HOST=ohpc137pbsop-sms.soft.fujitsu.com
SLURM_JOB_NAME=printenv
SLURM_JOB_CPUS_PER_NODE=6
SLURM_NTASKS=6
SLURM_NPROCS=6
SLURM_JOB_ID=276
SLURM_JOBID=276
SLURM_STEP_ID=0
SLURM_STEPID=0
SLURM_NNODES=1
SLURM_JOB_NUM_NODES=1
SLURM_NODELIST=ohpc137pbsop-c001
SLURM_JOB_PARTITION=regnodes
SLURM_TASKS_PER_NODE=6
SLURM_SRUN_COMM_PORT=37316
SLURM_JOB_ACCOUNT=test
SLURM_JOB_QOS=normal
SLURM_WORKING_CLUSTER=linux:ohpc137pbsop-sms:6817:8448
SLURM_JOB_NODELIST=ohpc137pbsop-c001
SLURM_STEP_NODELIST=ohpc137pbsop-c001
SLURM_STEP_NUM_NODES=1
SLURM_STEP_NUM_TASKS=6
SLURM_STEP_TASKS_PER_NODE=6
SLURM_STEP_LAUNCHER_PORT=37316
SLURM_SRUN_COMM_HOST=172.16.20.104
SLURM_TOPOLOGY_ADDR=ohpc137pbsop-c001
SLURM_TOPOLOGY_ADDR_PATTERN=node
SLURM_CPUS_ON_NODE=6
SLURM_CPU_BIND=verbose,cores
SLURM_CPU_BIND_LIST=
SLURM_CPU_BIND_TYPE=cores
SLURM_CPU_BIND_VERBOSE=verbose
SLURM_TASK_PID=130988
SLURM_NODEID=0
SLURM_PROCID=2
SLURM_LOCALID=2
SLURM_LAUNCH_NODE_IPADDR=172.16.20.104
SLURM_GTIDS=0,1,2,3,4,5
SLURM_CHECKPOINT_IMAGE_DIR=/var/slurm/checkpoint
SLURM_JOB_UID=1001
SLURM_JOB_USER=test
SLURM_JOB_GID=1001
SLURMD_NODENAME=ohpc137pbsop-c001
SLURM_PRIO_PROCESS=0
SLURM_UMASK=0002
SLURM_CLUSTER_NAME=linux
SLURM_SUBMIT_DIR=/home/test
SLURM_SUBMIT_HOST=ohpc137pbsop-sms.soft.fujitsu.com
SLURM_JOB_NAME=printenv
SLURM_JOB_CPUS_PER_NODE=6
SLURM_NTASKS=6
SLURM_NPROCS=6
SLURM_JOB_ID=276
SLURM_JOBID=276
SLURM_STEP_ID=0
SLURM_STEPID=0
SLURM_NNODES=1
SLURM_JOB_NUM_NODES=1
SLURM_NODELIST=ohpc137pbsop-c001
SLURM_JOB_PARTITION=regnodes
SLURM_TASKS_PER_NODE=6
SLURM_SRUN_COMM_PORT=37316
SLURM_JOB_ACCOUNT=test
SLURM_JOB_QOS=normal
SLURM_WORKING_CLUSTER=linux:ohpc137pbsop-sms:6817:8448
SLURM_JOB_NODELIST=ohpc137pbsop-c001
SLURM_STEP_NODELIST=ohpc137pbsop-c001
SLURM_STEP_NUM_NODES=1
SLURM_STEP_NUM_TASKS=6
SLURM_STEP_TASKS_PER_NODE=6
SLURM_STEP_LAUNCHER_PORT=37316
SLURM_SRUN_COMM_HOST=172.16.20.104
SLURM_TOPOLOGY_ADDR=ohpc137pbsop-c001
SLURM_TOPOLOGY_ADDR_PATTERN=node
SLURM_CPUS_ON_NODE=6
SLURM_CPU_BIND=verbose,cores
SLURM_CPU_BIND_LIST=
SLURM_CPU_BIND_TYPE=cores
SLURM_CPU_BIND_VERBOSE=verbose
SLURM_TASK_PID=130990
SLURM_NODEID=0
SLURM_PROCID=4
SLURM_LOCALID=4
SLURM_LAUNCH_NODE_IPADDR=172.16.20.104
SLURM_GTIDS=0,1,2,3,4,5
SLURM_CHECKPOINT_IMAGE_DIR=/var/slurm/checkpoint
SLURM_JOB_UID=1001
SLURM_JOB_USER=test
SLURM_JOB_GID=1001
SLURMD_NODENAME=ohpc137pbsop-c001
SLURM_PRIO_PROCESS=0
SLURM_UMASK=0002
SLURM_CLUSTER_NAME=linux
SLURM_SUBMIT_DIR=/home/test
SLURM_SUBMIT_HOST=ohpc137pbsop-sms.soft.fujitsu.com
SLURM_JOB_NAME=printenv
SLURM_JOB_CPUS_PER_NODE=6
SLURM_NTASKS=6
SLURM_NPROCS=6
SLURM_JOB_ID=276
SLURM_JOBID=276
SLURM_STEP_ID=0
SLURM_STEPID=0
SLURM_NNODES=1
SLURM_JOB_NUM_NODES=1
SLURM_NODELIST=ohpc137pbsop-c001
SLURM_JOB_PARTITION=regnodes
SLURM_TASKS_PER_NODE=6
SLURM_SRUN_COMM_PORT=37316
SLURM_JOB_ACCOUNT=test
SLURM_JOB_QOS=normal
SLURM_WORKING_CLUSTER=linux:ohpc137pbsop-sms:6817:8448
SLURM_JOB_NODELIST=ohpc137pbsop-c001
SLURM_STEP_NODELIST=ohpc137pbsop-c001
SLURM_STEP_NUM_NODES=1
SLURM_STEP_NUM_TASKS=6
SLURM_STEP_TASKS_PER_NODE=6
SLURM_STEP_LAUNCHER_PORT=37316
SLURM_SRUN_COMM_HOST=172.16.20.104
SLURM_TOPOLOGY_ADDR=ohpc137pbsop-c001
SLURM_TOPOLOGY_ADDR_PATTERN=node
SLURM_CPUS_ON_NODE=6
SLURM_CPU_BIND=verbose,cores
SLURM_CPU_BIND_LIST=
SLURM_CPU_BIND_TYPE=cores
SLURM_CPU_BIND_VERBOSE=verbose
SLURM_TASK_PID=130987
SLURM_NODEID=0
SLURM_PROCID=1
SLURM_LOCALID=1
SLURM_LAUNCH_NODE_IPADDR=172.16.20.104
SLURM_GTIDS=0,1,2,3,4,5
SLURM_CHECKPOINT_IMAGE_DIR=/var/slurm/checkpoint
SLURM_JOB_UID=1001
SLURM_JOB_USER=test
SLURM_JOB_GID=1001
SLURMD_NODENAME=ohpc137pbsop-c001
SLURM_PRIO_PROCESS=0
SLURM_UMASK=0002
SLURM_CLUSTER_NAME=linux
SLURM_SUBMIT_DIR=/home/test
SLURM_SUBMIT_HOST=ohpc137pbsop-sms.soft.fujitsu.com
SLURM_JOB_NAME=printenv
SLURM_JOB_CPUS_PER_NODE=6
SLURM_NTASKS=6
SLURM_NPROCS=6
SLURM_JOB_ID=276
SLURM_JOBID=276
SLURM_STEP_ID=0
SLURM_STEPID=0
SLURM_NNODES=1
SLURM_JOB_NUM_NODES=1
SLURM_NODELIST=ohpc137pbsop-c001
SLURM_JOB_PARTITION=regnodes
SLURM_TASKS_PER_NODE=6
SLURM_SRUN_COMM_PORT=37316
SLURM_JOB_ACCOUNT=test
SLURM_JOB_QOS=normal
SLURM_WORKING_CLUSTER=linux:ohpc137pbsop-sms:6817:8448
SLURM_JOB_NODELIST=ohpc137pbsop-c001
SLURM_STEP_NODELIST=ohpc137pbsop-c001
SLURM_STEP_NUM_NODES=1
SLURM_STEP_NUM_TASKS=6
SLURM_STEP_TASKS_PER_NODE=6
SLURM_STEP_LAUNCHER_PORT=37316
SLURM_SRUN_COMM_HOST=172.16.20.104
SLURM_TOPOLOGY_ADDR=ohpc137pbsop-c001
SLURM_TOPOLOGY_ADDR_PATTERN=node
SLURM_CPUS_ON_NODE=6
SLURM_CPU_BIND=verbose,cores
SLURM_CPU_BIND_LIST=
SLURM_CPU_BIND_TYPE=cores
SLURM_CPU_BIND_VERBOSE=verbose
SLURM_TASK_PID=130989
SLURM_NODEID=0
SLURM_PROCID=3
SLURM_LOCALID=3
SLURM_LAUNCH_NODE_IPADDR=172.16.20.104
SLURM_GTIDS=0,1,2,3,4,5
SLURM_CHECKPOINT_IMAGE_DIR=/var/slurm/checkpoint
SLURM_JOB_UID=1001
SLURM_JOB_USER=test
SLURM_JOB_GID=1001
SLURMD_NODENAME=ohpc137pbsop-c001
SLURM_PRIO_PROCESS=0
SLURM_UMASK=0002
SLURM_CLUSTER_NAME=linux
SLURM_SUBMIT_DIR=/home/test
SLURM_SUBMIT_HOST=ohpc137pbsop-sms.soft.fujitsu.com
SLURM_JOB_NAME=printenv
SLURM_JOB_CPUS_PER_NODE=6
SLURM_NTASKS=6
SLURM_NPROCS=6
SLURM_JOB_ID=276
SLURM_JOBID=276
SLURM_STEP_ID=0
SLURM_STEPID=0
SLURM_NNODES=1
SLURM_JOB_NUM_NODES=1
SLURM_NODELIST=ohpc137pbsop-c001
SLURM_JOB_PARTITION=regnodes
SLURM_TASKS_PER_NODE=6
SLURM_SRUN_COMM_PORT=37316
SLURM_JOB_ACCOUNT=test
SLURM_JOB_QOS=normal
SLURM_WORKING_CLUSTER=linux:ohpc137pbsop-sms:6817:8448
SLURM_JOB_NODELIST=ohpc137pbsop-c001
SLURM_STEP_NODELIST=ohpc137pbsop-c001
SLURM_STEP_NUM_NODES=1
SLURM_STEP_NUM_TASKS=6
SLURM_STEP_TASKS_PER_NODE=6
SLURM_STEP_LAUNCHER_PORT=37316
SLURM_SRUN_COMM_HOST=172.16.20.104
SLURM_TOPOLOGY_ADDR=ohpc137pbsop-c001
SLURM_TOPOLOGY_ADDR_PATTERN=node
SLURM_CPUS_ON_NODE=6
SLURM_CPU_BIND=verbose,cores
SLURM_CPU_BIND_LIST=
SLURM_CPU_BIND_TYPE=cores
SLURM_CPU_BIND_VERBOSE=verbose
SLURM_TASK_PID=130991
SLURM_NODEID=0
SLURM_PROCID=5
SLURM_LOCALID=5
SLURM_LAUNCH_NODE_IPADDR=172.16.20.104
SLURM_GTIDS=0,1,2,3,4,5
SLURM_CHECKPOINT_IMAGE_DIR=/var/slurm/checkpoint
SLURM_JOB_UID=1001
SLURM_JOB_USER=test
SLURM_JOB_GID=1001
SLURMD_NODENAME=ohpc137pbsop-c001
[test@ohpc137pbsop-sms ~]$



[test@ohpc137pbsop-sms ~]$ srun --nodes=1-3 lscpu
Architecture:          x86_64
CPU op-mode(s):        32-bit, 64-bit
Byte Order:            Little Endian
CPU(s):                48
On-line CPU(s) list:   0-47
Thread(s) per core:    2
Core(s) per socket:    12
Socket(s):             2
NUMA node(s):          4
Vendor ID:             GenuineIntel
CPU family:            6
Model:                 85
Model name:            Intel(R) Xeon(R) Gold 6126 CPU @ 2.60GHz
Stepping:              4
CPU MHz:               999.914
CPU max MHz:           3700.0000
CPU min MHz:           1000.0000
BogoMIPS:              5200.00
Virtualization:        VT-x
L1d cache:             32K
L1i cache:             32K
L2 cache:              1024K
L3 cache:              19712K
NUMA node0 CPU(s):     0-2,5-7,24-26,29-31
NUMA node1 CPU(s):     3,4,8-11,27,28,32-35
NUMA node2 CPU(s):     12-15,19,20,36-39,43,44
NUMA node3 CPU(s):     16-18,21-23,40-42,45-47
Flags:                 fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca 
cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx 
pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl 
xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 monitor ds_cpl 
vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic 
movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 
3dnowprefetch epb cat_l3 cdp_l3 intel_ppin intel_pt ssbd mba ibrs ibpb stibp 
tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 hle avx2 smep 
bmi2 erms invpcid rtm cqm mpx rdt_a avx512f avx512dq rdseed adx smap clflushopt 
clwb avx512cd avx512bw avx512vl xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc 
cqm_mbm_total cqm_mbm_local dtherm ida arat pln pts hwp hwp_act_window hwp_epp 
hwp_pkg_req pku ospke spec_ctrl intel_stibp flush_l1d
Architecture:          x86_64
CPU op-mode(s):        32-bit, 64-bit
Byte Order:            Little Endian
CPU(s):                48
On-line CPU(s) list:   0-47
Thread(s) per core:    2
Core(s) per socket:    12
Socket(s):             2
NUMA node(s):          4
Vendor ID:             GenuineIntel
CPU family:            6
Model:                 85
Model name:            Intel(R) Xeon(R) Gold 6126 CPU @ 2.60GHz
Stepping:              4
CPU MHz:               1000.073
CPU max MHz:           3700.0000
CPU min MHz:           1000.0000
BogoMIPS:              5200.00
Virtualization:        VT-x
L1d cache:             32K
L1i cache:             32K
L2 cache:              1024K
L3 cache:              19712K
NUMA node0 CPU(s):     0-2,5-7,24-26,29-31
NUMA node1 CPU(s):     3,4,8-11,27,28,32-35
NUMA node2 CPU(s):     12-14,18-20,36-38,42-44
NUMA node3 CPU(s):     15-17,21-23,39-41,45-47
Flags:                 fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca 
cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx 
pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl 
xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 monitor ds_cpl 
vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic 
movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 
3dnowprefetch epb cat_l3 cdp_l3 intel_ppin intel_pt ssbd mba ibrs ibpb stibp 
tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 hle avx2 smep 
bmi2 erms invpcid rtm cqm mpx rdt_a avx512f avx512dq rdseed adx smap clflushopt 
clwb avx512cd avx512bw avx512vl xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc 
cqm_mbm_total cqm_mbm_local dtherm ida arat pln pts hwp hwp_act_window hwp_epp 
hwp_pkg_req pku ospke spec_ctrl intel_stibp flush_l1d
Architecture:          x86_64
CPU op-mode(s):        32-bit, 64-bit
Byte Order:            Little Endian
CPU(s):                48
On-line CPU(s) list:   0-47
Thread(s) per core:    2
Core(s) per socket:    12
Socket(s):             2
NUMA node(s):          4
Vendor ID:             GenuineIntel
CPU family:            6
Model:                 85
Model name:            Intel(R) Xeon(R) Gold 6126 CPU @ 2.60GHz
Stepping:              4
CPU MHz:               1617.700
CPU max MHz:           3700.0000
CPU min MHz:           1000.0000
BogoMIPS:              5200.00
Virtualization:        VT-x
L1d cache:             32K
L1i cache:             32K
L2 cache:              1024K
L3 cache:              19712K
NUMA node0 CPU(s):     0-3,7,8,24-27,31,32
NUMA node1 CPU(s):     4-6,9-11,28-30,33-35
NUMA node2 CPU(s):     12-14,18-20,36-38,42-44
NUMA node3 CPU(s):     15-17,21-23,39-41,45-47
Flags:                 fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca 
cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx 
pdpe1gb rdtscp lm constant_tsc art arch_perfmon pebs bts rep_good nopl 
xtopology nonstop_tsc aperfmperf eagerfpu pni pclmulqdq dtes64 monitor ds_cpl 
vmx smx est tm2 ssse3 sdbg fma cx16 xtpr pdcm pcid dca sse4_1 sse4_2 x2apic 
movbe popcnt tsc_deadline_timer aes xsave avx f16c rdrand lahf_lm abm 
3dnowprefetch epb cat_l3 cdp_l3 intel_ppin intel_pt ssbd mba ibrs ibpb stibp 
tpr_shadow vnmi flexpriority ept vpid fsgsbase tsc_adjust bmi1 hle avx2 smep 
bmi2 erms invpcid rtm cqm mpx rdt_a avx512f avx512dq rdseed adx smap clflushopt 
clwb avx512cd avx512bw avx512vl xsaveopt xsavec xgetbv1 cqm_llc cqm_occup_llc 
cqm_mbm_total cqm_mbm_local dtherm ida arat pln pts hwp hwp_act_window hwp_epp 
hwp_pkg_req pku ospke spec_ctrl intel_stibp flush_l1d
[test@ohpc137pbsop-sms ~]$


[root@ohpc137pbsop-sms ~]# scontrol show config
Configuration data as of 2019-11-20T17:19:50
AccountingStorageBackupHost = (null)
AccountingStorageEnforce = none
AccountingStorageHost   = localhost
AccountingStorageLoc    = N/A
AccountingStoragePort   = 6819
AccountingStorageTRES   = cpu,mem,energy,node,billing,fs/disk,vmem,pages
AccountingStorageType   = accounting_storage/slurmdbd
AccountingStorageUser   = N/A
AccountingStoreJobComment = Yes
AcctGatherEnergyType    = acct_gather_energy/none
AcctGatherFilesystemType = acct_gather_filesystem/none
AcctGatherInterconnectType = acct_gather_interconnect/none
AcctGatherNodeFreq      = 0 sec
AcctGatherProfileType   = acct_gather_profile/none
AllowSpecResourcesUsage = 0
AuthInfo                = (null)
AuthType                = auth/munge
BatchStartTimeout       = 10 sec
BOOT_TIME               = 2019-11-20T17:19:40
BurstBufferType         = (null)
CheckpointType          = checkpoint/none
ClusterName             = linux
CommunicationParameters = (null)
CompleteWait            = 0 sec
CoreSpecPlugin          = core_spec/none
CpuFreqDef              = Unknown
CpuFreqGovernors        = Performance,OnDemand,UserSpace
CryptoType              = crypto/munge
DebugFlags              = (null)
DefMemPerNode           = UNLIMITED
DisableRootJobs         = No
EioTimeout              = 60
EnforcePartLimits       = NO
Epilog                  = (null)
EpilogMsgTime           = 2000 usec
EpilogSlurmctld         = (null)
ExtSensorsType          = ext_sensors/none
ExtSensorsFreq          = 0 sec
FastSchedule            = 1
FederationParameters    = (null)
FirstJobId              = 1
GetEnvTimeout           = 2 sec
GresTypes               = (null)
GroupUpdateForce        = 1
GroupUpdateTime         = 600 sec
HASH_VAL                = Match
HealthCheckInterval     = 0 sec
HealthCheckNodeState    = ANY
HealthCheckProgram      = (null)
InactiveLimit           = 0 sec
JobAcctGatherFrequency  = 30
JobAcctGatherType       = jobacct_gather/linux
JobAcctGatherParams     = (null)
JobCheckpointDir        = /var/slurm/checkpoint
JobCompHost             = localhost
JobCompLoc              = /var/log/slurm_jobcomp.log
JobCompPort             = 0
JobCompType             = jobcomp/none
JobCompUser             = root
JobContainerType        = job_container/none
JobCredentialPrivateKey = (null)
JobCredentialPublicCertificate = (null)
JobDefaults             = (null)
JobFileAppend           = 0
JobRequeue              = 1
JobSubmitPlugins        = (null)
KeepAliveTime           = SYSTEM_DEFAULT
KillOnBadExit           = 0
KillWait                = 30 sec
LaunchParameters        = (null)
LaunchType              = launch/slurm
Layouts                 =
Licenses                = (null)
LicensesUsed            = (null)
LogTimeFormat           = iso8601_ms
MailDomain              = (null)
MailProg                = /bin/mail
MaxArraySize            = 1001
MaxJobCount             = 10000
MaxJobId                = 67043328
MaxMemPerNode           = UNLIMITED
MaxStepCount            = 40000
MaxTasksPerNode         = 512
MCSPlugin               = mcs/none
MCSParameters           = (null)
MemLimitEnforce         = No
MessageTimeout          = 10 sec
MinJobAge               = 300 sec
MpiDefault              = none
MpiParams               = (null)
MsgAggregationParams    = (null)
NEXT_JOB_ID             = 271
NodeFeaturesPlugins     = (null)
OverTimeLimit           = 0 min
PluginDir               = /usr/lib64/slurm
PlugStackConfig         = /etc/slurm/plugstack.conf
PowerParameters         = (null)
PowerPlugin             =
PreemptMode             = OFF
PreemptType             = preempt/none
PriorityParameters      = (null)
PriorityType            = priority/basic
PrivateData             = none
ProctrackType           = proctrack/cgroup
Prolog                  = (null)
PrologEpilogTimeout     = 65534
PrologSlurmctld         = (null)
PrologFlags             = (null)
PropagatePrioProcess    = 0
PropagateResourceLimits = ALL
PropagateResourceLimitsExcept = (null)
RebootProgram           = (null)
ReconfigFlags           = (null)
RequeueExit             = (null)
RequeueExitHold         = (null)
ResumeFailProgram       = (null)
ResumeProgram           = (null)
ResumeRate              = 300 nodes/min
ResumeTimeout           = 60 sec
ResvEpilog              = (null)
ResvOverRun             = 0 min
ResvProlog              = (null)
ReturnToService         = 1
RoutePlugin             = route/default
SallocDefaultCommand    = (null)
SbcastParameters        = (null)
SchedulerParameters     = (null)
SchedulerTimeSlice      = 30 sec
SchedulerType           = sched/backfill
SelectType              = select/cons_res
SelectTypeParameters    = CR_CORE
SlurmUser               = slurm(202)
SlurmctldAddr           = (null)
SlurmctldDebug          = info
SlurmctldHost[0]        = ohpc137pbsop-sms
SlurmctldLogFile        = /var/log/slurmctld.log
SlurmctldPort           = 6817
SlurmctldSyslogDebug    = unknown
SlurmctldPrimaryOffProg = (null)
SlurmctldPrimaryOnProg  = (null)
SlurmctldTimeout        = 300 sec
SlurmctldParameters     = (null)
SlurmdDebug             = info
SlurmdLogFile           = /var/log/slurmd.log
SlurmdParameters        = (null)
SlurmdPidFile           = /var/run/slurmd.pid
SlurmdPort              = 6818
SlurmdSpoolDir          = /var/spool/slurm/d
SlurmdSyslogDebug       = unknown
SlurmdTimeout           = 300 sec
SlurmdUser              = root(0)
SlurmSchedLogFile       = (null)
SlurmSchedLogLevel      = 0
SlurmctldPidFile        = /var/run/slurmctld.pid
SlurmctldPlugstack      = (null)
SLURM_CONF              = /etc/slurm/slurm.conf
SLURM_VERSION           = 18.08.6
SrunEpilog              = (null)
SrunPortRange           = 0-0
SrunProlog              = (null)
StateSaveLocation       = /var/spool/slurm/ctld
SuspendExcNodes         = (null)
SuspendExcParts         = (null)
SuspendProgram          = (null)
SuspendRate             = 60 nodes/min
SuspendTime             = NONE
SuspendTimeout          = 30 sec
SwitchType              = switch/none
TaskEpilog              = (null)
TaskPlugin              = task/cgroup
TaskPluginParam         = (null type)
TaskProlog              = (null)
TCPTimeout              = 2 sec
TmpFS                   = /tmp
TopologyParam           = (null)
TopologyPlugin          = topology/none
TrackWCKey              = No
TreeWidth               = 50
UsePam                  = 0
UnkillableStepProgram   = (null)
UnkillableStepTimeout   = 60 sec
VSizeFactor             = 0 percent
WaitTime                = 0 sec
X11Parameters           = (null)

Cgroup Support Configuration:
AllowedDevicesFile      = /etc/slurm/cgroup_allowed_devices_file.conf
AllowedKmemSpace        = (null)
AllowedRAMSpace         = 100.0%
AllowedSwapSpace        = 0.0%
CgroupAutomount         = yes
CgroupMountpoint        = /cgroup
ConstrainCores          = yes
ConstrainDevices        = no
ConstrainKmemSpace      = no
ConstrainRAMSpace       = yes
ConstrainSwapSpace      = no
MaxKmemPercent          = 100.0%
MaxRAMPercent           = 100.0%
MaxSwapPercent          = 100.0%
MemLimitThreshold       = 100.0%
MemoryLimitEnforcement  = no
MemorySwappiness        = (null)
MinKmemSpace            = 30 MB
MinRAMSpace             = 30 MB
TaskAffinity            = yes

Slurmctld(primary) at ohpc137pbsop-sms is UP
[root@ohpc137pbsop-sms ~]#

Reply via email to