NOTICE: This version of the NSF Unidata web site (archive.unidata.ucar.edu) is no longer being updated.
Current content can be found at unidata.ucar.edu.
To learn about what's going on, see About the Archive Site.
Russ, >Date: Tue, 11 Sep 2001 16:35:41 -0600 >From: Russ Rew <russ@xxxxxxxxxxxxxxxx> >To: Konrad Hinsen <hinsen@xxxxxxxxxxxxxxx> >Subject: Re: performance degrades with filesize The above message contained the following: > Specifically, the following C program: > > http://www.unidata.ucar.edu/packages/netcdf/jg1.c > > accepts a single command-line argument for how many records to write, > creates a netCDF file with the same structure as John Galbraith's > Python script, and prints the time required to append each additional > batch of 5000 records to the initially empty file. > > The times are nearly constant even when the file has grown to 1.2 > Gbytes: I modified the program I wrote to use nc_put_varm_float rather than nc_put_vara_float. Performance still didn't degrade so the problem doesn't appear to depend on which specific output function is invoked. Regards, Steve Emmerson <http://www.unidata.ucar.edu>
netcdfgroup
archives: