Hi Timo,

Thanks for your help. Actually, to keep it simple, i am using my laptop for both server and client. So, i think there is no need to open ssh tunnel in this case and also hash will not be an issue.

Steps:

1 - installed Paraview 4.3.1 source under Mac OS 10.10 by enabling Python + OpenMPI + Development files and ParaView works fine.
2 - i compiled ParaViewCatalystExampleCode without any problem
3 - i run server: mpirun -np 2 ./pvserver (on port 11111)
4 - i run ParaView and connect to server and Catalyst (on port 22222) using menus 5 - i go to CxxMultiPieceExample directory and i run the code: mpirun -np 2 ./CxxMultiPieceExample SampleScripts/feslicescript.py

in this case, i modified SampleScripts/feslicescript.py as coprocessor.EnableLiveVisualization(True) to enable live visualization. In this case, the code is running but i am getting

Invalid output port : pqPipelineSource::getOutputPort( 0 ). Available number of output ports: 0

error under ParaView and also server crashes with following error,

ERROR: In /Users/turuncu/Qsync/progs/paraview-4.3.1/src/VTK/Parallel/Core/vtkSocketCommunicator.cxx, line 812
vtkSocketCommunicator (0x7f993cf20b80): Could not receive tag. 1

[epsilon:07510] *** Process received signal ***
[epsilon:07510] Signal: Segmentation fault: 11 (11)
[epsilon:07510] Signal code: Address not mapped (1)
[epsilon:07510] Failing at address: 0x30
[epsilon:07510] [ 0] 0 libsystem_platform.dylib 0x00007fff851f1f1a _sigtramp + 26
[epsilon:07510] [ 1] 0   ??? 0x00007f993b8578a2 0x0 + 140296105326754
[epsilon:07510] [ 2] 0 libvtkCommonSystem-pv4.3.1.dylib 0x000000010fd5c9d9 _ZN9vtkSocket4SendEPKvi + 41 [epsilon:07510] [ 3] 0 libvtkParallelCore-pv4.3.1.dylib 0x000000010f313803 _ZN21vtkSocketCommunicator10SendTaggedEPKviiiPKc + 67 [epsilon:07510] [ 4] 0 libvtkParallelCore-pv4.3.1.dylib 0x000000010f3131a1 _ZN21vtkSocketCommunicator13SendVoidArrayEPKvxiii + 4625 [epsilon:07510] [ 5] 0 libvtkPVClientServerCoreCore-pv4.3.1.dylib 0x000000010d5738e2 _ZN15vtkCommunicator4SendEPKcxii + 98 [epsilon:07510] [ 6] 0 libvtkPVClientServerCoreCore-pv4.3.1.dylib 0x000000010d56fa4d _ZN25vtkMultiProcessController4SendEPKcxii + 77 [epsilon:07510] [ 7] 0 libvtkPVClientServerCoreCore-pv4.3.1.dylib 0x000000010d56eca7 _ZN20vtkPVProgressHandler14RefreshMessageEPKc + 151 [epsilon:07510] [ 8] 0 libvtkPVClientServerCoreCore-pv4.3.1.dylib 0x000000010d56d99b _ZN20vtkPVProgressHandler14OnMessageEventEP9vtkObjectmPv + 91 [epsilon:07510] [ 9] 0 libvtkPVClientServerCoreCore-pv4.3.1.dylib 0x000000010d5719b8 _ZN9vtkObject22vtkClassMemberCallbackI20vtkPVProgressHandlerEclEPS_mPv + 344 [epsilon:07510] [10] 0 libvtkCommonCore-pv4.3.1.dylib 0x000000011056c7d2 _ZN24vtkObjectCommandInternal7ExecuteEP9vtkObjectmPv + 98 [epsilon:07510] [11] 0 libvtkCommonCore-pv4.3.1.dylib 0x0000000110569a7b _ZN16vtkSubjectHelper11InvokeEventEmPvP9vtkObject + 4731 [epsilon:07510] [12] 0 libvtkCommonCore-pv4.3.1.dylib 0x000000011056a052 _ZN9vtkObject11InvokeEventEmPv + 66 [epsilon:07510] [13] 0 libvtkCommonCore-pv4.3.1.dylib 0x0000000110571a7b _ZN15vtkOutputWindow11DisplayTextEPKc + 219 [epsilon:07510] [14] 0 libvtkCommonCore-pv4.3.1.dylib 0x0000000110571ac6 _ZN15vtkOutputWindow16DisplayErrorTextEPKc + 54 [epsilon:07510] [15] 0 libvtkCommonCore-pv4.3.1.dylib 0x000000011057165c _Z31vtkOutputWindowDisplayErrorTextPKc + 44 [epsilon:07510] [16] 0 libvtkParallelCore-pv4.3.1.dylib 0x000000010f315619 _ZN21vtkSocketCommunicator13ReceiveTaggedEPviiiPKc + 761 [epsilon:07510] [17] 0 libvtkParallelCore-pv4.3.1.dylib 0x000000010f315196 _ZN21vtkSocketCommunicator16ReceiveVoidArrayEPvxiii + 4486 [epsilon:07510] [18] 0 libvtkParallelCore-pv4.3.1.dylib 0x000000010f2d88a2 _ZN15vtkCommunicator7ReceiveEPhxii + 98 [epsilon:07510] [19] 0 libvtkParallelCore-pv4.3.1.dylib 0x000000010f2e91a1 _ZN25vtkMultiProcessController11ProcessRMIsEii + 193 [epsilon:07510] [20] 0 libvtkPVClientServerCoreCore-pv4.3.1.dylib 0x000000010d5bb89d _ZN26vtkTCPNetworkAccessManager21ProcessEventsInternalEmb + 2717 [epsilon:07510] [21] 0 libvtkPVClientServerCoreCore-pv4.3.1.dylib 0x000000010d5bba72 _ZN26vtkTCPNetworkAccessManager13ProcessEventsEm + 34 [epsilon:07510] [22] 0 pvserver 0x000000010ba83caa _ZL8RealMainiPPcN16vtkProcessModule12ProcessTypesE + 730
[epsilon:07510] [23] 0   pvserver 0x000000010ba839b7 main + 39
[epsilon:07510] [24] 0   libdyld.dylib 0x00007fff8cbe75c9 start + 1
[epsilon:07510] *** End of error message ***
--------------------------------------------------------------------------
mpirun noticed that process rank 0 with PID 7510 on node epsilon exited on signal 11 (Segmentation fault: 11).
--------------------------------------------------------------------------

So, i am unable to run Catalyst tests for now. They produces output *.vtm files (using writers in the pipeline) if i disable live visualization. Please let me know, if you have any suggestion.

Regards,

--ufuk

On 20/05/15 10:31, Timo Oster wrote:
Hi,

I have indeed run into the problem with "Could not receive tag. 1". In
my case I had to make sure that the client and server versions were
exactly the same. When connecting, the vtkSocketCommunicators on client
and server exchange a version hash that must match. This hash is
generated while building ParaView. The file that is generated can be
found in your build directory under
VTK/Parallel/Core/vtkSocketCommunicatorHash.h. Check that your client
and server have the same hash value in this file.

I am assuming that you have correctly set up an ssh tunnel between
client and server to enable the connection (ssh -L 11111:localhost:11111
user@server). The steps I take to run a in-situ case on a single remote
workstation are listed in my previous mail.

BTW I am using the "trunk" version of ParaView from the git repository,
which is close to version 4.3.1, and I'm running Linux on both client
and server (different distributions).

Regards,

Timo

On 05/19/2015 06:57 PM, u.utku.turunco...@be.itu.edu.tr wrote:
Hi Timo,

Have you ever tried to test Catalyst examples in your case.

https://github.com/Kitware/ParaViewCatalystExampleCode

I am also trying to run Catalyst examples but i could not. In my case, i
am able to install Paraview (4.3.1) from source under Mac OS 10.10 and the
Catalys examples but when i try to run the examples i am getting error
like,

ERROR: In
/Users/turuncu/Qsync/progs/paraview-4.3.1/src/VTK/Parallel/Core/vtkSocketCommunicator.cxx,
line 812
vtkSocketCommunicator (0x7f86a52658e0): Could not receive tag. 1

or

ERROR: In
/Users/turuncu/Qsync/progs/paraview-4.3.1/src/VTK/Common/System/vtkSocket.cxx,
line 572
vtkClientSocket (0x7fcefc939fe0): Socket error in call to send. Broken pipe.

So, if you don't mind could you share the details of your work environment
(OS, Paraview version, etc.) and the steps to run the in-situ example. So,
that could also help me to find the source of the problem.

Regards,

--ufuk

Hi all,

in an effort to enable live visualization of our simulation code, I have
written a Catalyst adaptor for it. The live visualization is working
great when the ParaView client runs on the same machine as the
simulation, even when the simulation runs in parallel using mpi.

Now I want to do live visualization of a simulation running on a remote
cluster. I am able to get this to work for the simulation running on the
login node of the cluster:

1. Tunnel an ssh connection of port 11111 to the cluster:
     ssh -L 11111:localhost:11111 server
2. In the shell that opens, start a paraview server (in the background
with '&')
3. Run the simulation with mpirun (simulation runs only on login node)
4. Start my local ParaView client and connect to localhost:11111 to
connect to the remote pvserver through the ssh tunnel
5. In the client, connect to Catalyst (port 22222)
6. A provider for the simulation is created and the live visualization
works

Now I want to do the same for simulations started on the remote cluster
via the batch job system. In this scenario, the parallel processes of
the simulation will run on different (randomly chosen) nodes on the
cluster. How do I go about getting a connection from my local client to
the Catalyst instances running on those nodes?

I imagine I will need to set up ssh tunnels from the nodes to the login
node where the pvserver is running. I've tried adding a ssh tunnel line
to the job script that is executed when the batch job starts. I've tried
forwarding and reverse-forwarding port 22222 (ssh -L and ssh -R) to no
avail. The best I get is "ERROR: In
/.../ParaView_git/VTK/Common/System/vtkSocket.cxx, line 206
vtkServerSocket (0x18e0930): Socket error in call to bind. Address
already in use."

My knowledge in networking and ssh is limited, so any pointers to how I
would go about this are greatly appreciated.

Regards,

Timo
_______________________________________________
Powered by www.kitware.com

Visit other Kitware open-source projects at
http://www.kitware.com/opensource/opensource.html

Please keep messages on-topic and check the ParaView Wiki at:
http://paraview.org/Wiki/ParaView

Search the list archives at: http://markmail.org/search/?q=ParaView

Follow this link to subscribe/unsubscribe:
http://public.kitware.com/mailman/listinfo/paraview


_______________________________________________
Powered by www.kitware.com

Visit other Kitware open-source projects at 
http://www.kitware.com/opensource/opensource.html

Please keep messages on-topic and check the ParaView Wiki at: 
http://paraview.org/Wiki/ParaView

Search the list archives at: http://markmail.org/search/?q=ParaView

Follow this link to subscribe/unsubscribe:
http://public.kitware.com/mailman/listinfo/paraview

Reply via email to