If your are trying to create a communicator containing all node local processes then use MPI_Comm_split_type.
> On Aug 3, 2018, at 12:24 PM, Diego Avesani <diego.aves...@gmail.com> wrote: > > Deal all, > probably I have found the error. > Let's me check. Probably I have not properly set-up colors. > > Thanks a lot, > I hope that you have not lost too much time for me, > I will let you know If that was the problem. > > Thanks again > > Diego > > >> On 3 August 2018 at 19:57, Diego Avesani <diego.aves...@gmail.com> wrote: >> Dear R, Dear all, >> >> I do not know. >> I have isolated the issues. It seem that I have some problem with: >> CALL >> MPI_COMM_SPLIT(MPI_COMM_WORLD,colorl,MPIworld%rank,MPI_LOCAL_COMM,MPIworld%iErr) >> CALL MPI_COMM_RANK(MPI_LOCAL_COMM, MPIlocal%rank,MPIlocal%iErr) >> CALL MPI_COMM_SIZE(MPI_LOCAL_COMM, MPIlocal%nCPU,MPIlocal%iErr) >> >> openMPI seems not able to create properly MPIlocal%rank. >> >> what should be? a bug? >> >> thanks again >> >> Diego >> >> >>> On 3 August 2018 at 19:47, Ralph H Castain <r...@open-mpi.org> wrote: >>> Those two command lines look exactly the same to me - what am I missing? >>> >>> >>>> On Aug 3, 2018, at 10:23 AM, Diego Avesani <diego.aves...@gmail.com> wrote: >>>> >>>> Dear all, >>>> >>>> I am experiencing a strange error. >>>> >>>> In my code I use three group communications: >>>> MPI_COMM_WORLD >>>> MPI_MASTERS_COMM >>>> LOCAL_COMM >>>> >>>> which have in common some CPUs. >>>> >>>> when I run my code as >>>> mpirun -np 4 --oversubscribe ./MPIHyperStrem >>>> >>>> I have no problem, while when I run it as >>>> >>>> mpirun -np 4 --oversubscribe ./MPIHyperStrem >>>> >>>> sometimes it crushes and sometimes not. >>>> >>>> It seems that all is linked to >>>> CALL MPI_REDUCE(QTS(tstep,:), QTS(tstep,:), nNode, MPI_DOUBLE_PRECISION, >>>> MPI_SUM, 0, MPI_LOCAL_COMM, iErr) >>>> >>>> which works with in local. >>>> >>>> What do you think? Can you please suggestion some debug test? >>>> Is a problem related to local communications? >>>> >>>> Thanks >>>> >>>> >>>> >>>> Diego >>>> >>>> _______________________________________________ >>>> users mailing list >>>> users@lists.open-mpi.org >>>> https://lists.open-mpi.org/mailman/listinfo/users >>> >>> >>> _______________________________________________ >>> users mailing list >>> users@lists.open-mpi.org >>> https://lists.open-mpi.org/mailman/listinfo/users >> > > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://lists.open-mpi.org/mailman/listinfo/users
_______________________________________________ users mailing list users@lists.open-mpi.org https://lists.open-mpi.org/mailman/listinfo/users