NOTICE: This version of the NSF Unidata web site (archive.unidata.ucar.edu) is no longer being updated.
Current content can be found at unidata.ucar.edu.
To learn about what's going on, see About the Archive Site.
Hi Davide, > On May 2, 2020, at 1:06 PM, Wei-Keng Liao <wkliao@xxxxxxxxxxxxxxxx> wrote: > > The dump information shows there are actually 8 datasets in the file. > Below is the start offsets, sizes, and end offsets of individual datasets . > There is not much padding space in between the datasets. > According to this, your file is expected to be of size 16 GB. > > dataset name start offset size end offset > BS_K_linearized1 2,379 8,100,000,000 8,100,002,379 > BS_K_linearized2 12,150,006,475 3,127,549,440 15,277,555,915 Thanks Wei-kang for preparing this useful information. These two are netCDF dimensions which in HDF5 (netCDF-4) files are stored as HDF5 datasets. The above information indicate these HDF5 datasets are taking up file space although they should not have any actual data. The netCDF library only needs a specific value in the NAME attribute of these datasets. I, too, suggest creating your files with the latest version of the netCDF library as it may improve what kind of create dataset information it passes to the HDF5 library. -Aleksandar
netcdfgroup
archives: