Hi,

thanks for looking into this.

After looking further into this it came to mind that the jobs were submitted 
with --switches=1. As we have a max of 18 nodes per
switch the probability is high that there just wasn't a switch whose nodes were 
idle although enough nodes in total were.

So I think it had nothing to do with the GrpNodes or MaxNodes limitation for 
that account but simply that the restriction of
switches=1 wasn't satisfied.

It would be nice if the job reason would be a bit more detailed than it is 
right now, at least for the admin who has to answer to
the users.

Regards,

        Uwe


Am 13.03.2015 um 12:28 schrieb John Desantis:
> 
> Uwe,
> 
> I didn't see anything outstanding in the logs which would suggest an error.
> 
> Looking at the "production" account, are there any other active jobs
> or allocations (by users in that account) in place using one node
> which would technically make the free nodes stand at 17 versus 18?
> What happens if you update the account so that there is no limit on
> GrpNodes?
> 
> Have you checked on all free nodes which should be able to host the
> job for issues?  Do you have any reservations in the system which are
> preserving resources? How does the partition configuration look?  Are
> you specifying MaxNodes?
> 
> If I were in your shoes, I'd test the account by using -w and
> --exclusive with 'salloc' on each of the nodes listed in the debug
> output below.  I'd run it 18 times to test the GrpNodes limit.
> 
> This is all that I can think of right now.  I'll have another espresso
> soon enough and will reply if anything else comes to mind.  I hope
> this helps!
> 
> John DeSantis
> 
> 
> 
> 2015-03-12 4:59 GMT-04:00 Uwe Sauter <[email protected]>:
>>
>> No one able to give a hint?
>>
>> Am 10.03.2015 um 17:05 schrieb Uwe Sauter:
>>>
>>> Hi,
>>>
>>> I have an account "production" configured with limitations GrpNodes=18, 
>>> MaxNodes=18, MaxWall=7-00:00:00, an associated user with
>>> limitations MaxNodes=18, MaxWall=7-00:00:00 and a QoS with limitations 
>>> Priority=10, GraceTime=00:00:00, PreemtMode=cluster,
>>> Flags=DenyOnLimit, UsageFact0r=1.0, MinCPUs=1.
>>>
>>> This user submitted a job that is within those limitations:
>>>
>>> JobId=14115 JobName=XXX
>>>    UserId=XXX(XXX) GroupId=XXX(XXX)
>>>    Priority=1214 Nice=0 Account=production QOS=normal
>>>    JobState=PENDING Reason=Resources Dependency=(null)
>>>    Requeue=0 Restarts=0 BatchFlag=1 Reboot=0 ExitCode=0:0
>>>    RunTime=00:00:00 TimeLimit=4-00:00:00 TimeMin=N/A
>>>    SubmitTime=2015-03-10T13:08:56 EligibleTime=2015-03-10T13:08:56
>>>    StartTime=Unknown EndTime=Unknown
>>>    PreemptTime=None SuspendTime=None SecsPreSuspend=0
>>>    Partition=MyPartition AllocNode:Sid=frontend:15414
>>>    ReqNodeList=(null) ExcNodeList=(null)
>>>    NodeList=(null)
>>>    NumNodes=18-18 NumCPUs=180 CPUs/Task=1 ReqB:S:C:T=0:0:*:*
>>>    Socks/Node=* NtasksPerN:B:S:C=0:0:*:* CoreSpec=*
>>>    MinCPUsNode=1 MinMemoryNode=0 MinTmpDiskNode=0
>>>    Features=(null) Gres=(null) Reservation=(null)
>>>    Shared=0 Contiguous=0 Licenses=(null) Network=(null)
>>>    Command=(null)
>>>    WorkDir=/XXX
>>>    StdErr=/XXX/slurm-14115.out
>>>    StdIn=/dev/null
>>>    StdOut=/XXX/slurm-14115.out
>>>    Switches=1@1-00:00:00
>>>
>>> Submitting the same job with a lower node count (e.g. 17) immediately 
>>> starts the job on that account. There is a second job with
>>> lower priority for that account in the queue and enough free nodes in the 
>>> cluster that a 18 node job is able to run.
>>>
>>> How can I debug what's going on and get this job running? Turning on 
>>> scheduler debugging only shows:
>>>
>>> scheduler log:
>>> [2015-03-10T13:33:11.834] sched: JobId=14115. State=PENDING. 
>>> Reason=Resources. Priority=1000. Partition=MyPartition.
>>> [2015-03-10T13:33:11.834] sched: JobId=14116. State=PENDING. 
>>> Reason=Priority(Priority), Priority=1000, Partition=MyPartition.
>>>
>>>
>>> slurmctld log:
>>> [2015-03-10T15:01:42.259] Set DebugFlags to 
>>> Backfill,BackfillMap,SelectType,TraceJobs
>>> [2015-03-10T15:02:32.508] 
>>> SchedulerParameters=default_queue_depth=100,max_rpc_cnt=0,max_sched_time=4,partition_job_depth=0
>>> [2015-03-10T15:03:02.178] backfill: beginning
>>> [2015-03-10T15:03:02.178] =========================================
>>> [2015-03-10T15:03:02.178] Begin:2015-03-10T15:03:02 End:2015-03-11T15:03:02
>>> Nodes:n[502901,503001,503101,503201,503301,510301,510401,510501,510601,510701,510901,511001,511101,511201,511301,511601,511701,511801,511901,512001,512201,512301,512401,512501,512601,512901,513001,513101,513201,513301,513501,513601,513701,513801,513901,520301,520401,520601,520701,520901,521001,521101,521201,521301,521601,521701,521801,521901,522001,522201,522301,522401,522501,522601,522901,523001,523101,523201,523301,523501,523601,523701,523801,523901]
>>> [2015-03-10T15:03:02.178] =========================================
>>> [2015-03-10T15:03:02.178] backfill test for JobID=14115 Prio=1000 
>>> Partition=MyPartition
>>> [2015-03-10T15:03:02.178] Test job 14115 at 2015-03-10T15:03:02 on
>>> n[502901,503001,503101,503201,503301,510301,510401,510501,510601,510701,510901,511001,511101,511201,511301,511601,511701,511801,511901,512001,512201,512301,512401,512501,512601,512901,513001,513101,513201,513301,513501,513601,513701,513801,513901,520301,520401,520601,520701,520901,521001,521101,521201,521301,521601,521701,521801,521901,522001,522201,522301,522401,522501,522601,522901,523001,523101,523201,523301,523501,523601,523701,523801,523901]
>>> [2015-03-10T15:03:02.178] backfill test for JobID=14116 Prio=1000 
>>> Partition=MyPartition
>>> [2015-03-10T15:03:02.178] Test job 14116 at 2015-03-10T15:03:02 on
>>> n[502901,503001,503101,503201,503301,510301,510401,510501,510601,510701,510901,511001,511101,511201,511301,511601,511701,511801,511901,512001,512201,512301,512401,512501,512601,512901,513001,513101,513201,513301,513501,513601,513701,513801,513901,520301,520401,520601,520701,520901,521001,521101,521201,521301,521601,521701,521801,521901,522001,522201,522301,522401,522501,522601,522901,523001,523101,523201,523301,523501,523601,523701,523801,523901]
>>> [2015-03-10T15:03:02.178] backfill: reached end of job queue
>>> [2015-03-10T15:03:02.178] backfill: completed testing 2(2) jobs, usec=775
>>> [2015-03-10T15:04:24.991] Set DebugFlags to none
>>>
>>>
>>>
>>>
>>> Thanks,
>>>
>>>       Uwe
>> >

Reply via email to