CLASSIFICATION: UNCLASSIFIED

I've got a cluster with about 39 nodes, with 8 to 12 cores each. When I submit 
a job array job of say 15k, about 300 of those jobs start up across the 
cluster, but once those jobs complete, I only see one node's worth of jobs (say 
8) going at a time from then on and always on just one node, and I don't see 
the other nodes getting used at all. Would anyone have an idea as to why my 
other nodes don't continue to get jobs placed on them? Here are some pertinent 
settings from my slurm.conf

FastSchedule=0
SchedulerType=sched/builtin
SelectType=select/serial
SchedulerParameters=default_queue_depth=300
        
JobCompType=jobcomp/none
JobAcctGatherType=jobacct_gather/none

MaxJobCount=50000
TaskPlugin=task/none

InactiveLimit=0
KillWait=30
MinJobAge=2
SlurmctldTimeout=120
SlurmdTimeout=300
WaitTime=0

Thanks,
Tony

CLASSIFICATION: UNCLASSIFIED

Reply via email to