[Paraview] Output of ParallelUnstructuredGridWriter running on multiple nodes
Andy Bauer
andy.bauer at kitware.com
Thu Apr 25 19:43:05 EDT 2013
Hi,
The parallel XML data set writers all write independently of each other (no
MPI communication is needed in the writers). Because of this they can't
tell process 0 when writing it's meta-data file that it doesn't need to
write out meta data for them.
Andy
On Thu, Apr 25, 2013 at 4:30 PM, Hong Yi <hongyi at renci.org> wrote:
> I am running a simulation in-situ linked with ParaView CoProcessing
> library with a pipeline that includes “slice->Integrate Variables
> Filter->ParallelUnstructuredGridWriter” so that integrated velocity value
> for a slice at each time point can be written out while running the
> simulation in-situ. The output from this ParallelUnstructureGridWriter for
> each time step includes one pvtu file plus one vtu file from each running
> node, so that is a lot of output files for each time point especially when
> running on a large number of nodes in parallel. However, it looks like only
> the output vtu file from processing node 0 includes valid velocity range
> value while other output vtu files from other nodes include only
> blank/empty values. So I am wondering whether there is a good way to set up
> only outputting vtu file from processing node 0 to prevent output of those
> large number of files that don’t include valid/useful values. Any
> suggestions are very much appreciated. ****
>
> ** **
>
> Thanks,****
>
> ** **
>
> Hong****
>
> _______________________________________________
> Powered by www.kitware.com
>
> Visit other Kitware open-source projects at
> http://www.kitware.com/opensource/opensource.html
>
> Please keep messages on-topic and check the ParaView Wiki at:
> http://paraview.org/Wiki/ParaView
>
> Follow this link to subscribe/unsubscribe:
> http://www.paraview.org/mailman/listinfo/paraview
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://www.paraview.org/pipermail/paraview/attachments/20130425/e93aec56/attachment.htm>
More information about the ParaView
mailing list