From: Forafo S. <ppv...@gm...> - 2013-08-29 15:40:32
|
Hello All, I have some data in an HDF5 file that is created with PyTables. Occasionally, I update the data by reading in one of the tables and adding or deleting rows. Then, I create a new table containing the updated data, give it a random name, and let it reside in the same group where the old table resides. I flush the new table, then use the table.remove() (or Leaf.remove()) method to delete the old table and table.rename() method to rename the randomly-named new table to the same name as the old table. Problem: In a small sized table, the size of the hdf5 file doubles with the above process even when no new rows or other modifications are made (let's assume that the hdf5 file contains only this table). A ptdump indicates no presence of the old table. In a medium-sized table, the size of the hdf5 file rises substantially (20% or 30%) even when no new rows or columns are added. Do I understand the table.remove() right as completely deleting the table? Does it leave some residue that I should be aware of? All help is appreciated. Thanks, Premal |