[OMPI devel] OFI, destroy_vni_context(1137).......: OFI domain close failed (ofi_init.c:1137:destroy_vni_context:Device or resource busy)

2022-11-01 Thread mrlong via devel
Hi, teachers code: import mpi4py import time import numpy as np from mpi4py import MPI comm = MPI.COMM_WORLD rank = comm.Get_rank() print("rank",rank) if __name__ == '__main__':     if rank == 0:     mem = np.array([0], dtype='i')     win = MPI.Win.Create(mem, comm=comm)     else:    

Re: [OMPI devel] [EXTERNAL] [OMPI users] OFI, destroy_vni_context(1137).......: OFI domain close failed (ofi_init.c:1137:destroy_vni_context:Device or resource busy)

2022-11-01 Thread mrlong via devel
Thanks, what you said seems to be right, I just checked and solved it. It might be caused by a conflict between openmpi and mpich library. 在 2022/11/2 02:06, Pritchard Jr., Howard 写道: HI, You are using MPICH or a vendor derivative of MPICH.  You probably want to resend this email to the mpic

[OMPI devel] [LOG_CAT_ML] component basesmuma is not available but requested in hierarchy: basesmuma, basesmuma, ucx_p2p:basesmsocket, basesmuma, p2p

2022-11-07 Thread mrlong via devel
The execution of openmpi 5.0.0rc9 results in the following: (py3.9) [user@machine01 share]$  mpirun -n 2 python test.py [LOG_CAT_ML] component basesmuma is not available but requested in hierarchy: basesmuma,basesmuma,ucx_p2p:basesmsocket,basesmuma,p2p [LOG_CAT_ML] ml_discover_hierarchy exited

[OMPI devel] There are not enough slots available in the system to satisfy the 2, slots that were requested by the application

2022-11-07 Thread mrlong via devel
*Two machines, each with 64 cores. The contents of the hosts file are:* 192.168.180.48 slots=1 192.168.60.203 slots=1 *Why do you get the following error when running with openmpi 5.0.0rc9?* (py3.9) [user@machine01 share]$  mpirun -n 2 --machinefile hosts hostname --