PERSONAL/NONWORK // EXTERNAL I took a look through the archives, and I did not see an clear answer to the issue I was seeing, so I thought I would go ahead and ask.
I am having a cluster issue with SLURM and I hoped you might be able to help me out. I built a small test cluster to determine if it might meet some compute needs I have but seem to keep running into an issue where SLURM is restricting MATLAB to using a single CPU regardless of how many we request. During testing I found the following: When I login into a MATLAB interactive session and run "feature numcores" I get the following: [cid:image002.jpg@01D48D35.180C4F30] Which is correct, as I have 14 cores and they are all available. However when I go into SLURM and request a MATLAB interactive session and run the same command on the same computer: [cid:image004.png@01D48D35.180C4F30] So, what I understand is that my cgroups settings in SLURM are restricting MATLAB to a single core. Is that correct? Also, how do I fix this? Here is my cgroups.conf ### # # Slurm cgroup support configuration file # # See man slurm.conf and man cgroup.conf for further # information on cgroup configuration parameters #-- ######################################################## # W A R N I N G: This file is managed by Puppet # # - - - - - - - changes are likely to be overwritten # ######################################################## ####################### CgroupAutomount=yes ####################### # testing -=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=- #ConstrainCores=no #ConstrainRAMSpace=no #ConstrainSwapSpace=no # testing -=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=- ConstrainCores=yes ConstrainRAMSpace=yes ConstrainSwapSpace=yes # ConstrainDevices=no AllowedSwapSpace=0 MaxRAMPercent=100 MaxSwapPercent=100 ##MinRAMSpace=30 # TASK/CGROUP PLUGIN # Constrain the job cgroup RAM to this percentage of the allocated memory. #AllowedRAMSpace=10 AllowedRamSpace=100 # TaskAffinity=<yes|no> # If configured to "yes" then set a default task affinity to bind each # step task to a subset of the allocated cores using # sched_setaffinity. The default value is "no". Note: This feature # requires the Portable Hardware Locality (hwloc) library to be # installed. TaskAffinity=yes # MemorySwappiness=<number> # Configure the kernel's priority for swapping out anonymous pages (such as program data) # verses file cache pages for the job cgroup. Valid values are between 0 and 100, inclusive. A # value of 0 prevents the kernel from swapping out program data. A value of 100 gives equal # priorioty to swapping out file cache or anonymous pages. If not set, then the kernel's default # swappiness value will be used. Either ConstrainRAMSpace or ConstrainSwapSpace must # be set to yes in order for this parameter to be applied. MemorySwappiness=0 ##################################################################################### # If compute nodes mount Lustre or NFS file systems, it may be a good idea to # # configure cgroup.conf with: # # ConstrainKmemSpace=no # # # # From <https://wiki.fysik.dtu.dk/niflheim/Slurm_configuration#activating-cgroups> # ##################################################################################### ConstrainKmemSpace=no # ######################################################## # W A R N I N G: This file is managed by Puppet # # - - - - - - - changes are likely to be overwritten # ######################################################## Thanks, Wes (A slurm neophyte)