Hi Timo,

There's a couple of steps that need to be added when connecting to a
Catalyst-enabled simulation that's been run through a batch job on a
cluster. This "should" be the same as just running pvserver through a batch
job and connecting to that. The basics are that there's an
executable/script that runs on the login node that the batch job connects
to. Your GUI connects to the login node and has the ports forwarded
properly through the executable/script to your pvserver. Now if your
pvserver is running on the login node then it will have to use the above
connection mechanism to connect to your Catalyst-enabled simulation. There
may be some emails or something on the wiki that has information on the
connection executable/script but I haven't tried any of that in several
years so I'm not sure where it would be. Maybe someone else will chime in
that's done this more recently than me.

Good luck and let us know how it goes.

Andy

On Mon, May 18, 2015 at 8:48 AM, Timo Oster <timo.os...@ovgu.de> wrote:

> Hi all,
>
> in an effort to enable live visualization of our simulation code, I have
> written a Catalyst adaptor for it. The live visualization is working
> great when the ParaView client runs on the same machine as the
> simulation, even when the simulation runs in parallel using mpi.
>
> Now I want to do live visualization of a simulation running on a remote
> cluster. I am able to get this to work for the simulation running on the
> login node of the cluster:
>
> 1. Tunnel an ssh connection of port 11111 to the cluster:
>     ssh -L 11111:localhost:11111 server
> 2. In the shell that opens, start a paraview server (in the background
> with '&')
> 3. Run the simulation with mpirun (simulation runs only on login node)
> 4. Start my local ParaView client and connect to localhost:11111 to
> connect to the remote pvserver through the ssh tunnel
> 5. In the client, connect to Catalyst (port 22222)
> 6. A provider for the simulation is created and the live visualization
> works
>
> Now I want to do the same for simulations started on the remote cluster
> via the batch job system. In this scenario, the parallel processes of
> the simulation will run on different (randomly chosen) nodes on the
> cluster. How do I go about getting a connection from my local client to
> the Catalyst instances running on those nodes?
>
> I imagine I will need to set up ssh tunnels from the nodes to the login
> node where the pvserver is running. I've tried adding a ssh tunnel line
> to the job script that is executed when the batch job starts. I've tried
> forwarding and reverse-forwarding port 22222 (ssh -L and ssh -R) to no
> avail. The best I get is "ERROR: In
> /.../ParaView_git/VTK/Common/System/vtkSocket.cxx, line 206
> vtkServerSocket (0x18e0930): Socket error in call to bind. Address
> already in use."
>
> My knowledge in networking and ssh is limited, so any pointers to how I
> would go about this are greatly appreciated.
>
> Regards,
>
> Timo
> _______________________________________________
> Powered by www.kitware.com
>
> Visit other Kitware open-source projects at
> http://www.kitware.com/opensource/opensource.html
>
> Please keep messages on-topic and check the ParaView Wiki at:
> http://paraview.org/Wiki/ParaView
>
> Search the list archives at: http://markmail.org/search/?q=ParaView
>
> Follow this link to subscribe/unsubscribe:
> http://public.kitware.com/mailman/listinfo/paraview
>
_______________________________________________
Powered by www.kitware.com

Visit other Kitware open-source projects at 
http://www.kitware.com/opensource/opensource.html

Please keep messages on-topic and check the ParaView Wiki at: 
http://paraview.org/Wiki/ParaView

Search the list archives at: http://markmail.org/search/?q=ParaView

Follow this link to subscribe/unsubscribe:
http://public.kitware.com/mailman/listinfo/paraview

Reply via email to