A Wednesday 22 September 2010 22:04:53 David E. Sallis escrigué:
> I have a table in an HDF5 file consisting of 9 columns and just over
> 6000 rows, and an application which performs updates on these table
> rows.  The application runs hourly and performs updates to the table
> during each run.  No new table rows are added during a run.  I
> perform updates to the table by using row.update() inside a
> table.where() iterator loop.
> 
> I have noticed that after each application run the size of the file
> increases significantly, and over time the file size balloons from
> just over 21 MB to well over 750 MB, with no new data being added,
> just updated.
> 
> h5repack() run on this file will restore it to its original size with
> no loss of data.
> 
> My questions are:
> 
> 1) What causes the file size increase and
> 2) is there anything I can do to prevent it?
> 
> I am using PyTables 2.1.1, HDF5 1.8.3, Python 2.6 under Linux RedHat
> 5.

Hmm, sounds like a bug somewhere.  Could you try with a recent version 
of HDF5?  If that does not help, could you please send me a small script 
reproducing that?

-- 
Francesc Alted

------------------------------------------------------------------------------
Start uncovering the many advantages of virtual appliances
and start using them to simplify application deployment and
accelerate your shift to cloud computing.
http://p.sf.net/sfu/novell-sfdev2dev
_______________________________________________
Pytables-users mailing list
Pytables-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/pytables-users

Reply via email to