hi!

  I recently installed paraview 3.12 with offscreen mesa on a SGI cluster
with intel compilers and SGI MPT. Using the same version of paraview on my
local computer I recorded a script in the qt version using python trace. I
was able to execute the same script just fine on the cluster on a similar
but larger case using pvpython. However I'm unable to use pvbatch. First
off the cluster forces me to use mpirun to run pvbatch, saying that all MPI
applications must be launched with mpirun. Even when I do this

mpirun -np 1 pvbatch --use-offscreen-rendering saveImage.py
MPI: MPI_COMM_WORLD rank 0 has terminated without calling MPI_Finalize()
MPI: aborting job
MPI: Received signal 9


a termination is all I get. Is something wrong the way I'm using pvbatch?
I'll be working on datasets that are over 20 million cells. I hope to use
multiple processors if that will help me speed up visualization. Please
note that I'm not using a server. I just intend to submit the visualization
as a job on the same cluster where I run the simulation.

ganesh

Attachment: saveImage.py
Description: Binary data

_______________________________________________
Powered by www.kitware.com

Visit other Kitware open-source projects at 
http://www.kitware.com/opensource/opensource.html

Please keep messages on-topic and check the ParaView Wiki at: 
http://paraview.org/Wiki/ParaView

Follow this link to subscribe/unsubscribe:
http://www.paraview.org/mailman/listinfo/paraview

Reply via email to