Sorry for the confusion. i'll try to explain more on my situation.
the first example show that the pvserver is "launchable" when it only
involves one machine as the pvserver.
i could connect to it using any binary version of  paraview client from any
machine.

at the second example, i'm trying to run a parallel server using 2 machines.

added the mpi bin directory PATH in my bashrc file and came out with
following konsole lines:


> [EMAIL PROTECTED]:~/installer/ParaView-3.4.0_build/bin> mpirun -hostfile
> ../../openmpi-1.2.8_build/etc/quadcore_frodo -np 6 ./pvserver
> --use-offscreen-rendering
> **Password:
> **[192.168.0.10][0,1,1][btl_tcp_endpoint.c:572:mca_btl_tcp_endpoint_complete_connect]
> connect() failed with errno=110
> **[192.168.0.10][0,1,3][btl_tcp_endpoint.c:572:mca_btl_tcp_endpoint_complete_connect]
> connect() failed with errno=110*
>

then the machine just fails to give any response.
when pressed with "control c" the following shows:


> mpirun: killing job...
>
*mpirun noticed that job rank 0 with PID 20731 on node 192.168.0.10 exited
> on signal 15 (Terminated).
> 5 additional processes aborted (not shown)*


 Any ideas what going on?

thanks,



On Fri, Nov 7, 2008 at 12:08 AM, John M. Patchett <[EMAIL PROTECTED]> wrote:

>
> On Nov 6, 2008, at 8:34 AM, Moreland, Kenneth wrote:
>
> I am a bit confused by your question.  Are you saying the first example
> works and the second does not?  I am also confused by what you are trying to
> set up.  Are you trying to run the server on one of your machines and the
> client on the other, or are you trying to run a parallel server using both
> machines?
>
> In the second example, it looks like your OpenMPI install is messed up.  In
> the 4th line (1st error), it complains about not finding orted, which is a
> daemon that the OpenMPI version of mpirun uses to launch processes.  It does
> not look like pvserver is being run at all.
>
>
> I think Ken is correct.  It would appear that you don't have your bash
> environment set up correctly.  You need to have in your .profile or .bashrc
> or .something that is read when OpenMPI launches your job, the path to the
> mpi bin directory.
> the LD_LIBRARY_PATH needs to be set to your OpenMPI lib directory.  You can
> test that the PATH component is  set correctly for your type of login by
> running a command like:
> ssh 192.168.0.16 which mpirun
> The LD_LIBRARY_PATH will become obvious once your pvserver executable is
> actually launched.
> (The man page for bash on your system should explain what order the dot
> files are read in for interactive and non interactive shells...)
>
> At the very very end of this mail I'm pasting the show output of a very
> (very) old openmpi environment modules ...
> Hope this helps,
> -John.
>
>
>
> I cannot claim to have a lot of experience in setting up MPI installs,
> especially ones on heterogeneous nodes like you are setting up.  My advice
> would be to first make sure that you can launch a job "locally" on each
> machine, then try to launch a job only one the opposite computer (i.e.
> launch only on the 32 bit machine from the 64 bit machine and vice versa),
> and then finally launch a single job on both.
>
> -Ken
>
>
> On 11/6/08 1:32 AM, "yewyong" <[EMAIL PROTECTED]> wrote:
>
> Hi all,
>
> i managed to compile the paraview 3.4.0 in both a 64bit dual core machine
> and a 32bit quad core machine.
> having trouble starting the pvserver when linking 2 machines. Below are the
> konsole lines..
>
>  [EMAIL PROTECTED]:~/installer/ParaView-3.4.0_build/bin>
> ../../openmpi-1.2.8_build/bin/mpirun -hostfile
> ../../openmpi-1.2.8_build/etc/frodo -np 2 ./pvserver
> --use-offscreen-rendering
> Listen on port: 11111
> Waiting for client...
> ^Cmpirun: killing job...
> *
> *mpirun noticed that job rank 0 with PID 27000 on node 192.168.0.16 <
> http://192.168.0.16>  exited on signal 15 (Terminated).
> 1 additional process aborted (not shown)
> *
>
>
> everything looks good when there is no linking of 2 machines. (this is the
> 64bit dual core machine)
> but,
>
> [EMAIL PROTECTED]:~/installer/ParaView-3.4.0_build/bin>
> ../../openmpi-1.2.8_build/bin/mpirun -hostfile
> ../../openmpi-1.2.8_build/etc/frodo_quadcore -np 6 ./pvserver
> --use-offscreen-rendering
> Password:
> bash: orted: command not found
> [frodo:27009] [0,0,0] ORTE_ERROR_LOG: Timeout in file
> base/pls_base_orted_cmds.c at line 275
> [frodo:27009] [0,0,0] ORTE_ERROR_LOG: Timeout in file pls_rsh_module.c at
> line 1166
> [frodo:27009] [0,0,0] ORTE_ERROR_LOG: Timeout in file errmgr_hnp.c at line
> 90
> [frodo:27009] ERROR: A daemon on node 192.168.0.10 <http://192.168.0.10>
>  failed to start as expected.
> [frodo:27009] ERROR: There may be more information available from
> [frodo:27009] ERROR: the remote shell (see above).
> [frodo:27009] ERROR: The daemon exited unexpectedly with status 127.
> [frodo:27009] [0,0,0] ORTE_ERROR_LOG: Timeout in file
> base/pls_base_orted_cmds.c at line 188
> [frodo:27009] [0,0,0] ORTE_ERROR_LOG: Timeout in file pls_rsh_module.c at
> line 1198
> --------------------------------------------------------------------------
> mpirun was unable to cleanly terminate the daemons for this job. Returned
> value Timeout instead of ORTE_SUCCESS.
> --------------------------------------------------------------------------
> *
>
>
> i can't start the pvserver when involving 2 machines.
> 192.168.0.16 <http://192.168.0.16>  = 64bit dual core
> 192.168.0.10 <http://192.168.0.10>  = 32bit quad core
>
> Anyone faced this before?
> Thanks in advance for helping.
>
> Cheers,
> yewyong
> VRC UM
>
>
>
>
>    ****      Kenneth Moreland
>     ***      Sandia National Laboratories
> ***********
> *** *** ***  email: [EMAIL PROTECTED]
> **  ***  **  phone: (505) 844-8919
>     ***      fax:   (505) 845-0833
>
> _______________________________________________
> ParaView mailing list
> ParaView@paraview.org
> http://www.paraview.org/mailman/listinfo/paraview
>
>
>
> [dq001 ~]$module show openmpi
> -------------------------------------------------------------------
> /usr/local/Modules/modulefiles/openmpi/1.0.1curr:
>
> prepend-path     PATH
> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/bin
> prepend-path     MANPATH
> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/man
> prepend-path     LD_LIBRARY_PATH
> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib
> prepend-path     LIBRARY_PATH /usr/mellanox/lib
> prepend-path     PATH /usr/mellanox/bin
> prepend-path     CPLUS_INCLUDE_PATH
> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/include
> prepend-path     C_INCLUDE_PATH
> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/include
> prepend-path     LIBRARY_PATH
> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib
> prepend-path     LIBRARY_PATH /usr/mellanox/lib
> setenv           MPI_ROOT
> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install
> setenv           LDLIBS -lmpi
> setenv           PV_MPI_INCLUDE_PATH
> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/include/openmpi/ompi
> setenv           PV_MPI_LIBRARY
> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib/libmpi.a
> setenv           PV_MPI_EXTRA_LIBRARY
> /usr/lib64/libmtl_common.so;/usr/lib64/libvapi.so;/usr/lib64/libmpga.so;/usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib/libmpi_cxx.a;/usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib/libopal.a;/usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib/liborte.a;/usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib/libmpi.a;/usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib/libmpi_cxx.a;/usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib/libopal.a;/usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/lib/liborte.a;/usr/lib64/libutil.a
> /usr/lib64/libc.a
> setenv           PV_MPI_NUMPROC_FLAG -np
> setenv           PV_MPIRUN
> /usr/local/packages/openmpi/current/installs/mpi_get_ompi_1.0/mpi_install_infiniband/install/bin/mpirun
> -------------------------------------------------------------------
>
_______________________________________________
ParaView mailing list
ParaView@paraview.org
http://www.paraview.org/mailman/listinfo/paraview

Reply via email to