The addresses are different, but the shared objects seem to be the same. --Adam
On Apr 17, 2012, at 12:17 PM, John Patchett wrote: > Sometimes, depending on the process launcher being used by MPI you don't get > the same environment in your current shell as you get from the mpiexeced > environment, since interactive shells behave differently than non-interactive > shells in regards to what dot files they read and/or don't read. > > you should try an > ssh ldd /home/dersh/projects/ParaView-bin/bin/pvserver > or > rsh ldd /home/dersh/projects/ParaView-bin/bin/pvserver > if you get the same output as > ldd /home/dersh/projects/ParaView-bin/bin/pvserver > I'm barking up the wrong tree. > > Good Luck, > --John. > > On Tue, Apr 17, 2012 at 1:02 PM, Utkarsh Ayachit > <utkarsh.ayac...@kitware.com> wrote: > Are there multiple version of MPI installed by any chance? > > Utkarsh > > On Tue, Apr 17, 2012 at 2:51 PM, Adam Dershowitz > <adershow...@exponent.com> wrote: > > Attached are the results of ldd libvtkParallel.so (looks to me like it is > > linking MPI stuff). And also CMakeCache.txt > > > > > > > > > > > > -----Original Message----- > > From: Utkarsh Ayachit [mailto:utkarsh.ayac...@kitware.com] > > Sent: Tue 4/17/2012 6:14 AM > > To: Adam Dershowitz > > Cc: paraview@paraview.org > > Subject: Re: [Paraview] MPI Socket in use > > > > Do you mind posting your CMakeCache.txt file? Also do a "ldd > > libvtkParallel.so". Let's verify that vtkParallel is linking against > > MPI libs. > > > > Utkarsh > > > > On Tue, Apr 17, 2012 at 12:26 AM, Adam Dershowitz, Ph.D., P.E. > > <adershow...@exponent.com> wrote: > >> I am sure that I use ccmake and changed PARAVIEW_USE_MPI to ON, (I also > >> enabled python and pointed to the open mpi compiler, so that it then filled > >> in most of the variables for MPI, as the note explains). Then I did > >> configure and generate. Finally I did make, and sudo make install. All > >> seemed to work fine, and paraview runs fine, with a single processor. > >> I even went so far and to make a new, empty directory, and rebuilt it > >> there. With the same results. > >> I have also tried to explicitly put in the correct path, just to make sure > >> that there isn't some other pvserver around: > >> mpirun -np 3 /home/dersh/projects/ParaView-bin/bin/pvserver > >> with exactly the same results. > >> I have gone back, and now I see that MPIEXEC_MAX_NUMPROCS and > >> VTK_MPI_MAX_NUMPROCS are set to 2. But, I get the same error if I try to > >> run it with -np 2, so clearly just using MPI is failing. A > >> > >> I realized that the problem is consistent with it not being built with > >> MPI, but I definitely set it on. Are there some other variables that have > >> to be set? Clearly something is not being built correctly, but I am not > >> sure what it is. > >> > >> Thanks, > >> > >> --Adam > >> > >> > >> > >> On Apr 16, 2012, at 7:09 PM, Utkarsh Ayachit wrote: > >> > >>> You may want to verify PARAVIEW_USE_MPI flag again and ensure that the > >>> pvserver you're running is indeed the one that has PARAVIEW_USE_MPI > >>> set to ON. The problem you're seeing is typical when ParaView not > >>> built with MPI. > >>> > >>> Utkarsh > >>> > >>> On Mon, Apr 16, 2012 at 8:10 PM, Adam Dershowitz > >>> <adershow...@exponent.com> wrote: > >>>> I just built paraview on an opensuse linux box. When I use the GUI, and > >>>> a > >>>> single connection it seems to work fine. But, if I try to use multiple > >>>> CPUs, or run with mpi, it fails. > >>>> I do have OpenMPI installed. > >>>> When I first started getting the error, I googled around some and found > >>>> that > >>>> maybe "make install" would help (I had been running it just in the build > >>>> directory). But I am getting the same error after installing. I also > >>>> added > >>>> my openmpi libraries to LD_LIBRARY_PATH (when I first tried to run it I > >>>> had > >>>> other errors about a shared library). > >>>> > >>>> I did build with PARAVIEW_USE_MPI set to on. It looks as though one > >>>> pvserver will run, but any additional ones finds a port in use. Clearly > >>>> something is not right about how MPI is being handled, but I am not sure > >>>> how > >>>> to fix it. > >>>> > >>>> > >>>> If I try mpirun, here is error I get: > >>>> > >>>> > >>>> mpiexec -np 3 pvserver > >>>> Waiting for client... > >>>> Connection URL: cs://cfd:11111 > >>>> Accepting connection(s): cfd:11111 > >>>> Waiting for client... > >>>> Connection URL: cs://cfd:11111 > >>>> ERROR: In > >>>> /home/dersh/projects/ParaView-3.14.1-Source/VTK/Common/vtkSocket.cxx, > >>>> line > >>>> 206 > >>>> vtkServerSocket (0xebd970): Socket error in call to bind. Address > >>>> already in > >>>> use. > >>>> > >>>> ERROR: In > >>>> > >>>> /home/dersh/projects/ParaView-3.14.1-Source/ParaViewCore/ClientServerCore/vtkTCPNetworkAccessManager.cxx, > >>>> line 343 > >>>> vtkTCPNetworkAccessManager (0x661800): Failed to set up server socket. > >>>> > >>>> Exiting... > >>>> Waiting for client... > >>>> Connection URL: cs://cfd:11111 > >>>> ERROR: In > >>>> /home/dersh/projects/ParaView-3.14.1-Source/VTK/Common/vtkSocket.cxx, > >>>> line > >>>> 206 > >>>> vtkServerSocket (0xebd970): Socket error in call to bind. Address > >>>> already in > >>>> use. > >>>> > >>>> ERROR: In > >>>> > >>>> /home/dersh/projects/ParaView-3.14.1-Source/ParaViewCore/ClientServerCore/vtkTCPNetworkAccessManager.cxx, > >>>> line 343 > >>>> vtkTCPNetworkAccessManager (0x661800): Failed to set up server socket. > >>>> > >>>> Exiting... > >>>> > >>>> ----------------------------------------------------------------------------- > >>>> It seems that [at least] one of the processes that was started with > >>>> mpirun did not invoke MPI_INIT before quitting (it is possible that > >>>> more than one process did not invoke MPI_INIT -- mpirun was only > >>>> notified of the first one, which was on node n0). > >>>> > >>>> mpirun can *only* be used with MPI programs (i.e., programs that > >>>> invoke MPI_INIT and MPI_FINALIZE). You can use the "lamexec" program > >>>> to run non-MPI programs over the lambooted nodes. > >>>> > >>>> ----------------------------------------------------------------------------- > >>>> mpirun failed with exit status 252 > >>>> > >>>> And suggestions would be greatly appreciated. > >>>> > >>>> thanks, > >>>> > >>>> --Adam > >>>> > >>>> > >>>> _______________________________________________ > >>>> Powered by www.kitware.com > >>>> > >>>> Visit other Kitware open-source projects at > >>>> http://www.kitware.com/opensource/opensource.html > >>>> > >>>> Please keep messages on-topic and check the ParaView Wiki at: > >>>> http://paraview.org/Wiki/ParaView > >>>> > >>>> Follow this link to subscribe/unsubscribe: > >>>> http://www.paraview.org/mailman/listinfo/paraview > >>>> > >> > > > _______________________________________________ > Powered by www.kitware.com > > Visit other Kitware open-source projects at > http://www.kitware.com/opensource/opensource.html > > Please keep messages on-topic and check the ParaView Wiki at: > http://paraview.org/Wiki/ParaView > > Follow this link to subscribe/unsubscribe: > http://www.paraview.org/mailman/listinfo/paraview >
_______________________________________________ Powered by www.kitware.com Visit other Kitware open-source projects at http://www.kitware.com/opensource/opensource.html Please keep messages on-topic and check the ParaView Wiki at: http://paraview.org/Wiki/ParaView Follow this link to subscribe/unsubscribe: http://www.paraview.org/mailman/listinfo/paraview