Hi David,

> The new xdmf library's treatment of the data structures that pertain to the 
> xml contents it much more efficient than the old version. 
> Also, at the ParaView level there you have two choices of how to read and 
> xdmf file with the new reader. 
> The "(Top Level Parition)" is meant for this case. It makes it so that that 
> every node opens its own child xdmf files. That way no memory is spent on the 
> xml structured for contents they are not responsible for the hdf5 data for.

That would be the perfect approach for my use case :-)
Could you point me to some documentation on how to use the new XDMF reader?
I did not yet find much on Google & Co.

I am using Paraview 4.2.0 for Linux x64. I also tried a Nightly-Build in 
November.
The XDMF-Files state the specification as 2.2 from 2003.

Is there anything special necessary to make Paraview use the new reader and 
provide the option for top-level-partitioning?
Or do I have to compile Paraview myself and enable some configure-flags?

Thanks a lot and happy holidays :-)
--
Dipl.-Phys. Karl-Ulrich Bamberg

Ludwig-Maximilians-Universität München
Arnold-Sommerfeld-Center (ASC)
Computational & Plasma Physics
Theresienstr. 37, D-80333 München

phone: +49 (0)89 2180 4577
fax: +49 (0)89 2180 99 4577
e-mail: [email protected]

Am 19.12.2014 um 23:23 schrieb David E DeMarle:

> Hi Karl,
> 
> Sorry for the lag in response.
> 
> We spent quite a bit of time looking at this issue when we were making the 
> xdmf3 reader.
> 
> The new xdmf library's treatment of the data structures that pertain to the 
> xml contents it much more efficient than the old version. 
> 
> Also, at the ParaView level there you have two choices of how to read and 
> xdmf file with the new reader. 
> 
> The "(Top Level Parition)" is meant for this case. It makes it so that that 
> every node opens its own child xdmf files. That way no memory is spent on the 
> xml structured for contents they are not responsible for the hdf5 data for.
> 
> hth
> 
> 
> David E DeMarle
> Kitware, Inc.
> R&D Engineer
> 21 Corporate Drive
> Clifton Park, NY 12065-8662
> Phone: 518-881-4909
> 
> On Wed, Dec 3, 2014 at 7:05 PM, Karl-Ulrich Bamberg 
> <[email protected]> wrote:
> Hi,
> 
> Paraview is a great and handy tool. I am now trying to scale it to large data 
> sets and have
> a question related to the parallel capabilities of the Paraview-XDMF-Reader:
> 
> We use a PIC-Code that devides its grid into blocks distributed over the 
> ranks.
> For the output every rank writes exactly one HDF5, collecting all local 
> blocks and all time steps.
> One XDMF per rank was written describing the data, as well as a central XDMF 
> including all the individual ones.
> 
> The hierarchy is 
>       central-xdmf:  spatial_collection->includes_for_rank_xdmf_files
>       rank_xdmf_files: 
> temporal_collection->spatial_collectio_within_ranks->grids
> 
> The size of the simulation is about 1024 ranks and 256 time steps but should 
> be increased.
> For these parameters we see (via top) a memory consumption of 16GB per 
> pvserver instance.
> Directly after opening of the file, so even before "apply". 
> 
> I guess that this is because all the pvserver instances read and parse the 
> XDMF file?
> One time the paths to the HDF5 files were wrong, the memory consumption was 
> the same.
> After "apply" there was than an error.
> 
> I tried "PV_USE_TRANSMIT=1" and also changed the grid hierarchy to only have:
> temporal_collection->spatial_collection->grids
> 
> This directly in one file, that was finally 1GB on disk and around 16GB in 
> memory with lxml2 via python.
> 
> But it was to no effort, still every instance was consuming around 16GB
> 
> Is there any chance that pvserver can parallelize on the top-level so every 
> pvserver instance only reads some of the "include" files.
> Or is there a different approach to store the grid-patches (all the same 
> resolution right now) in HDF5?
> 
> All suggestions are highly appreciated :-)
> 
> Thank you all very much for any support,
> Best regards
> --
> Dipl.-Phys. Karl-Ulrich Bamberg
> 
> Ludwig-Maximilians-Universität München
> Arnold-Sommerfeld-Center (ASC)
> Computational & Plasma Physics
> Theresienstr. 37, D-80333 München
> 
> phone: +49 (0)89 2180 4577
> fax: +49 (0)89 2180 99 4577
> e-mail: [email protected]
> 
> 
> _______________________________________________
> Powered by www.kitware.com
> 
> Visit other Kitware open-source projects at 
> http://www.kitware.com/opensource/opensource.html
> 
> Please keep messages on-topic and check the ParaView Wiki at: 
> http://paraview.org/Wiki/ParaView
> 
> Follow this link to subscribe/unsubscribe:
> http://public.kitware.com/mailman/listinfo/paraview
> 

Attachment: signature.asc
Description: Message signed with OpenPGP using GPGMail

_______________________________________________
Powered by www.kitware.com

Visit other Kitware open-source projects at 
http://www.kitware.com/opensource/opensource.html

Please keep messages on-topic and check the ParaView Wiki at: 
http://paraview.org/Wiki/ParaView

Follow this link to subscribe/unsubscribe:
http://public.kitware.com/mailman/listinfo/paraview

Reply via email to