NOTICE: This version of the NSF Unidata web site (archive.unidata.ucar.edu) is no longer being updated.
Current content can be found at unidata.ucar.edu.
To learn about what's going on, see About the Archive Site.
Hi, Before I throw myself into implementing this, has anyone else modified nccopy to algorithmically determine chunking specs for variables, or allowed this to be specified on a per variable basis? I’m trying the approach documented by Russ Rew here: http://www.unidata.ucar.edu/blogs/developer/en/entry/chunking_data_choosing_shapes and have a python algorithm working for my 3 and 4 dimensional variables. I had hoped to use nccopy and pass these chunking schemes to it, but as nccopy so succinctly describes in its own man page: "Note that nccopy requires variables that share a dimension to also share the chunk size associated with that dimension, but the programming interface has no such restriction. If you need to customize chunking for variables independently, you will need to use the library API in a custom utility program.” and I have 3D and 4D variables in the same file which share dimensions. I am thinking the easiest route may be for me to fork nccopy and implement this optimal chunking algorithm internally. Does this sound like a good idea? I was hoping someone else was already implementing something like this. Thanks Aidan
netcdfgroup
archives: