Hi,A few tasks with higher priority give way to tasks with lower priority, and I don't understand why.
I noticed that the hi-prio tasks require 4 or 8 x GPUs on a single node, while the bypassing tasks only use 1 x GPU, but I'm not sure if it's related. High-priority tasks have a specific value in the StartTime field, but regularly, this value is pushed back to a later time. It seems like after finishing a 1GPU task, Slurm immediately schedules another 1GPU task instead of waiting for the release of the remaining 3 or 7 GPUs for a higher-priority task. What can be wrong?
The tasks are being launched in the 'long' partition with QoS named long. PartitionName=long AllowGroups=ALL AllowAccounts=ALL AllowQos=ALL AllocNodes=ALL Default=NO QoS=longDefaultTime=2-00:00:00 DisableRootJobs=NO ExclusiveUser=NO GraceTime=0 Hidden=NO MaxNodes=UNLIMITED MaxTime=10-00:00:00 MinNodes=0 LLN=NO MaxCPUsPerNode=UNLIMITED
Nodes=dgx-[1-4],sr-[1-3]PriorityJobFactor=1 PriorityTier=10000 RootOnly=NO ReqResv=NO OverSubscribe=NO
OverTimeLimit=NONE PreemptMode=SUSPEND State=UP TotalCPUs=656 TotalNodes=7 SelectTypeParameters=NONE JobDefaults=(null) DefMemPerNode=UNLIMITED MaxMemPerNode=UNLIMITED TRES=cpu=656,mem=8255731M,node=7,billing=3474,gres/gpu=32,gres/gpu:a100=32 TRESBillingWeights=CPU=1,Mem=0.062G,GRES/gpu=72.458 Name GrpTRES ------ --------------------------- normal long cpu=450,gres/gpu=28,mem=5T Example of bypassed job with obscured sensitive data: $ scontrol show job 649800 JobId=649800 JobName=----train with motif UserId=XXXXXX(XXXX) GroupId=XXXXXX(XXXXX) MCS_label=N/A Priority=275000 Nice=0 Account=sfglab QOS=normal JobState=PENDING Reason=QOSGrpGRES Dependency=(null) Requeue=1 Restarts=0 BatchFlag=1 Reboot=0 ExitCode=0:0 RunTime=00:00:00 TimeLimit=03:00:00 TimeMin=N/A SubmitTime=2023-04-24T12:09:19 EligibleTime=2023-04-24T12:09:19 AccrueTime=2023-04-24T12:09:19 StartTime=2023-05-11T06:30:00 EndTime=2023-05-11T09:30:00 Deadline=N/ASuspendTime=None SecsPreSuspend=0 LastSchedEval=2023-05-09T13:16:46 Scheduler=Backfill:*
Partition=long AllocNode:Sid=0.0.0.0:379113 ReqNodeList=(null) ExcNodeList=(null) NodeList= NumNodes=1 NumCPUs=16 NumTasks=1 CPUs/Task=16 ReqB:S:C:T=0:0:*:* TRES=cpu=16,mem=32G,node=1,billing=16,gres/gpu=8 Socks/Node=* NtasksPerN:B:S:C=0:0:*:* CoreSpec=* MinCPUsNode=16 MinMemoryNode=32G MinTmpDiskNode=0 Features=(null) DelayBoot=00:00:00 OverSubscribe=OK Contiguous=0 Licenses=(null) Network=(null) Command=/XXXXX WorkDir=/XXXXX StdErr=/XXXXXX StdIn=/dev/null StdOut=/XXXXX Power= TresPerNode=gres:gpu:8 -- best regards | pozdrawiam serdecznie *Michał Kadlof*Head of the high performance computing center Kierownik ośrodka obliczeniowego HPC
Eden^N cluster administrator Administrator klastra obliczeniowego Eden^NStructural and Functional Genomics Laboratory Laboratorium Genomiki Strukturalnej i Funkcjonalnej Faculty of Mathematics and Computer Science Wydział Matematyki i Nauk Informacyjnych
Warsaw University of Technology Politechnika Warszawska
smime.p7s
Description: S/MIME Cryptographic Signature