NOTICE: This version of the NSF Unidata web site (archive.unidata.ucar.edu) is no longer being updated.
Current content can be found at unidata.ucar.edu.
To learn about what's going on, see About the Archive Site.
Hi, Something seems to have changed between Netcdf 4.3.1 and 4.3.2 where compression has stopped working for my application, which creates a number of 3 and 4-dimensional variables. One of the dimensions is unlimited. I'm using the 4.4.0 fortran (f90) interface. Ncdump -s shows the same attributes for both 4.3.1 and 4.3.2, for example: QC:_ChunkSizes = 1, 10, 40, 40 ; QC:_DeflateLevel = 2 ; QC:_Shuffle = "true" ; But the file size from using 4.3.2 is the same as when deflate is turned off. Very strange. I tested one of the check programs in nf_test with a larger array, and it seems to work as expected, however, with both 4.3.1 and 4.3.2 (i.e., it compresses successfully with both). I changed f90tst_vars4.f90 to write out a 4d variable with unlimited time dimension, but it does not reproduce the problem. We are also seeing a similar problem with the WRF model history files (even with just one time level), where it appears that some of the variables are getting compressed, but some are not (i.e., the files are little smaller (about 30%) than without deflate, but still substantially larger than when using an earlier version of netcdf, which are about 1/3 the non-deflated size). Has anybody else noticed this? The data write out fine and no errors are generated (or at least not detected). -- Ted __________________________________________________________ | Edward Mansell <ted.mansell@xxxxxxxx> | National Severe Storms Laboratory, Norman, OK. |-------------------------------------------------------------- | "The contents of this message are mine personally and | do not reflect any position of the U.S. Government or NOAA." |--------------------------------------------------------------
netcdfgroup
archives: