[Nexus] large overhead on small hdf5 files
Osborn, Raymond
rosborn at anl.gov
Wed Sep 24 16:43:40 BST 2014
There’s a StackOverflow issue at http://stackoverflow.com/questions/14332193/hdf5-storage-overhead that discusses this.
If you’re using NeXus and you have a large number of groups, perhaps this is a case where they should be consolidated - can you put all the metadata into a single NXparameters group? Or perhaps a gzipped XML file is the way to go. Or you just resign yourself to buying bigger hard drives.
Ray
On Sep 23, 2014, at 6:05 PM, Paul Kienzle <Paul.Kienzle at nist.gov> wrote:
> Anyone have any hints for reducing HDF5 file sizes?
>
> Using h5stat, I have 750k of metadata for 75k of raw data. The data is all small, but with 1300 data items in the file, the overhead adds up.
>
> I’ve been using the h5py python HDF5 wrapper to create the files, but even when copying them using napi they are still huge.
>
> Thanks in advance,
>
> - Paul
>
> Paul Kienzle
> paul.kienzle at nist.gov
>
>
> _______________________________________________
> NeXus mailing list
> NeXus at nexusformat.org
> http://lists.nexusformat.org/mailman/listinfo/nexus
--
Ray Osborn, Senior Scientist
Materials Science Division
Argonne National Laboratory
Argonne, IL 60439, USA
Phone: +1 (630) 252-9011
Email: ROsborn at anl.gov
More information about the NeXus
mailing list