If your are trying to create a communicator containing all node local processes 
then use MPI_Comm_split_type. 

> On Aug 3, 2018, at 12:24 PM, Diego Avesani <diego.aves...@gmail.com> wrote:
> 
> Deal all,
> probably I have found the error.
> Let's me check. Probably I have not properly set-up colors.
> 
> Thanks a lot,
> I hope that you have not lost too much time for me,
> I will let you know If that was the problem.
> 
> Thanks again
> 
> Diego
> 
> 
>> On 3 August 2018 at 19:57, Diego Avesani <diego.aves...@gmail.com> wrote:
>> Dear R, Dear all,
>> 
>> I do not know. 
>> I have isolated the issues. It seem that I have some problem with:
>>   CALL 
>> MPI_COMM_SPLIT(MPI_COMM_WORLD,colorl,MPIworld%rank,MPI_LOCAL_COMM,MPIworld%iErr)
>>   CALL MPI_COMM_RANK(MPI_LOCAL_COMM, MPIlocal%rank,MPIlocal%iErr)
>>   CALL MPI_COMM_SIZE(MPI_LOCAL_COMM, MPIlocal%nCPU,MPIlocal%iErr) 
>> 
>> openMPI seems not able to create properly MPIlocal%rank.
>> 
>> what should be? a bug?
>> 
>> thanks again
>> 
>> Diego
>> 
>> 
>>> On 3 August 2018 at 19:47, Ralph H Castain <r...@open-mpi.org> wrote:
>>> Those two command lines look exactly the same to me - what am I missing?
>>> 
>>> 
>>>> On Aug 3, 2018, at 10:23 AM, Diego Avesani <diego.aves...@gmail.com> wrote:
>>>> 
>>>> Dear all,
>>>> 
>>>> I am experiencing a strange error.
>>>> 
>>>> In my code I use three group communications:
>>>> MPI_COMM_WORLD
>>>> MPI_MASTERS_COMM
>>>> LOCAL_COMM
>>>> 
>>>> which have in common some CPUs.
>>>> 
>>>> when I run my code as 
>>>>  mpirun -np 4 --oversubscribe ./MPIHyperStrem
>>>> 
>>>> I have no problem, while when I run it as
>>>>  
>>>>  mpirun -np 4 --oversubscribe ./MPIHyperStrem
>>>> 
>>>> sometimes it crushes and sometimes not.
>>>> 
>>>> It seems that all is linked to 
>>>> CALL MPI_REDUCE(QTS(tstep,:), QTS(tstep,:), nNode, MPI_DOUBLE_PRECISION, 
>>>> MPI_SUM, 0, MPI_LOCAL_COMM, iErr)
>>>> 
>>>> which works with in local.
>>>> 
>>>> What do you think? Can you please suggestion some debug test?
>>>> Is a problem related to local communications?
>>>> 
>>>> Thanks
>>>> 
>>>> 
>>>> 
>>>> Diego
>>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> users@lists.open-mpi.org
>>>> https://lists.open-mpi.org/mailman/listinfo/users
>>> 
>>> 
>>> _______________________________________________
>>> users mailing list
>>> users@lists.open-mpi.org
>>> https://lists.open-mpi.org/mailman/listinfo/users
>> 
> 
> _______________________________________________
> users mailing list
> users@lists.open-mpi.org
> https://lists.open-mpi.org/mailman/listinfo/users
_______________________________________________
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

Reply via email to