Set rmaps_base_verbose=10 for debugging output 

Sent from my iPhone

> On Nov 1, 2018, at 9:31 AM, Adam LeBlanc <alebl...@iol.unh.edu> wrote:
> 
> The version by the way for Open-MPI is 3.1.2.
> 
> -Adam LeBlanc
> 
>> On Thu, Nov 1, 2018 at 12:05 PM Adam LeBlanc <alebl...@iol.unh.edu> wrote:
>> Hello,
>> 
>> I am an employee of the UNH InterOperability Lab, and we are in the process 
>> of testing OFED-4.17-RC1 for the OpenFabrics Alliance. We have purchased 
>> some new hardware that has one processor, and noticed an issue when running 
>> mpi jobs on nodes that do not have similar processor counts. If we launch 
>> the MPI job from a node that has 2 processors, it will fail and stating 
>> there are not enough resources and will not start the run, like so:
>> 
>> --------------------------------------------------------------------------
>> There are not enough slots available in the system to satisfy the 14 slots
>> that were requested by the application:
>>   IMB-MPI1
>> 
>> Either request fewer slots for your application, or make more slots available
>> for use.
>> --------------------------------------------------------------------------
>> 
>> If we launch the MPI job from the node with one processor, without changing 
>> the mpirun command at all, it runs as expected.
>> 
>> Here is the command being run:
>> 
>> mpirun --mca btl_openib_warn_no_device_params_found 0 --mca 
>> orte_base_help_aggregate 0 --mca btl openib,vader,self --mca pml ob1 --mca 
>> btl_openib_receive_queues P,65536,120,64,32 -hostfile 
>> /home/soesterreich/ce-mpi-hosts IMB-MPI1
>> 
>> Here is the hostfile being used:
>> 
>> farbauti-ce.ofa.iol.unh.edu slots=1
>> hyperion-ce.ofa.iol.unh.edu slots=1
>> io-ce.ofa.iol.unh.edu slots=1
>> jarnsaxa-ce.ofa.iol.unh.edu slots=1
>> rhea-ce.ofa.iol.unh.edu slots=1
>> tarqeq-ce.ofa.iol.unh.edu slots=1
>> tarvos-ce.ofa.iol.unh.edu slots=1
>> 
>> This seems like a bug and we would like some help to explain and fix what is 
>> happening. The IBTA plugfest saw similar behaviours, so this should be 
>> reproduceable.
>> 
>> Thanks,
>> Adam LeBlanc
> _______________________________________________
> users mailing list
> users@lists.open-mpi.org
> https://lists.open-mpi.org/mailman/listinfo/users
_______________________________________________
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

Reply via email to