NOTICE: This version of the NSF Unidata web site (archive.unidata.ucar.edu) is no longer being updated.
Current content can be found at unidata.ucar.edu.
To learn about what's going on, see About the Archive Site.
NOTE: The netcdf-hdf
mailing list is no longer active. The list archives are made available for historical reasons.
Hi Quincey, Thanks for the information. It's very helpful. I guess I will have to create a new file for every record instead. Regards, Eh Quincey Koziol wrote: > > On Nov 27, 2006, at 8:24 PM, Eh Tan wrote: > >> >> In the event of system crash, how can I prevent the file corruption and >> how can I minimize the loss of data? >> Should I flush the buffer after each output, or close the dataset after >> each output, or save each record in a new datagroup, or save each >> record >> in a new file? How much of data loss would I expect in the worst >> scenario (e.g., the system crashes during disk I/O)? > > > Generally, it's a good idea to call H5Fflush (or the equivalent > netCDF API call) after each major "phase" of writing to the file. > This will flush metadata changes out to the disk. However, it is > still possible that incremental changes may be made to the file as > metadata is evicted from the HDF5 internal caches that would create a > "corrupt" file if the rest of the changes don't make it into the > file. Flushing too often may create additional I/O though, so you'll > need to find a balance that's appropriate for your application. > > -- Eh Tan Staff Scientist Computational Infrastructure for Geodynamics 2750 E. Washington Blvd. Suite 210 Pasadena, CA 91107 (626) 395-1693 http://www.geodynamics.org ============================================================================== To unsubscribe netcdf-hdf, visit: http://www.unidata.ucar.edu/mailing-list-delete-form.html ==============================================================================
netcdf-hdf
archives: