Then the hang is curious. 

> On Sep 20, 2019, at 11:28 PM, Mills, Richard Tran <rtmi...@anl.gov> wrote:
> 
> Everything that Barry says about '--with-batch' is valid, but let me point 
> out one thing about Summit: You don't need "--with-batch" at all, because the 
> Summit login/compile nodes run the same hardware (minus the GPUs) and 
> software stack as the back-end compute nodes. This makes configuring and 
> building software far, far easier than we are used to on the big LCF 
> machines. I was actually shocked when I found this out -- I'd gotten so used 
> to struggling with cross-compilers, etc.
> 
> --Richard
> 
> On 9/20/19 9:28 PM, Smith, Barry F. wrote:
>>     --with-batch is still there and should be used in such circumstances. 
>> The difference is that --with-branch does not generate a program that you 
>> need to submit to the batch system before continuing the configure. Instead 
>> --with-batch guesses at and skips some of the tests (with clear warnings on 
>> how you can adjust the guesses).
>> 
>>      Regarding the hanging. This happens because the thread monitoring of 
>> configure started executables was removed years ago since it was slow and 
>> occasionally buggy (the default wait was an absurd 10 minutes too). Thus 
>> when configure tried to test an mpiexec that hung the test would hang.   
>> There is code in one of my branches I've been struggling to get into master 
>> for a long time that puts back the thread monitoring for this one call with 
>> a small timeout so you should never see this hang again.
>> 
>>   Barry
>> 
>>    We could be a little clever and have configure detect it is on a Cray or 
>> other batch system and automatically add the batch option. That would be a 
>> nice little feature for someone to add. Probably just a few lines of code. 
>>    
>> 
>> 
>>> On Sep 20, 2019, at 8:59 PM, Mills, Richard Tran via petsc-dev 
>>> <petsc-dev@mcs.anl.gov>
>>>  wrote:
>>> 
>>> Hi Junchao,
>>> 
>>> Glad you've found a workaround, but I don't know why you are hitting this 
>>> problem. The last time I built PETSc on Summit (just a couple days ago), I 
>>> didn't have this problem. I'm working from the example template that's in 
>>> the PETSc repo at config/examples/arch-olcf-summit-opt.py.
>>> 
>>> Can you point me to your configure script on Summit so I can try to 
>>> reproduce your problem?
>>> 
>>> --Richard
>>> 
>>> On 9/20/19 4:25 PM, Zhang, Junchao via petsc-dev wrote:
>>> 
>>>> Satish's trick --with-mpiexec=/bin/true solved the problem.  Thanks.
>>>> --Junchao Zhang
>>>> 
>>>> 
>>>> On Fri, Sep 20, 2019 at 3:50 PM Junchao Zhang 
>>>> <jczh...@mcs.anl.gov>
>>>>  wrote:
>>>> My configure hangs on Summit at
>>>>   TESTING: configureMPIEXEC from 
>>>> config.packages.MPI(config/BuildSystem/config/packages/MPI.py:170)
>>>> 
>>>> On the machine one has to use script to submit jobs. So why do we need 
>>>> configureMPIEXEC? Do I need to use --with-batch? I remember we removed 
>>>> that.
>>>> 
>>>> --Junchao Zhang
>>>> 
> 

Reply via email to