>> Computing the min & max on the fly is cheap, and approximating it is even >> cheaper, so why introduce the uncertainty? > >... but computing min & max on the fly can also be very expensive. >We have aggregated model output datasets where each variable is more >than 1TB!
Sure, I can see that that's useful metadata about the dataset, and that there's value in caching it somewhere. I just don't think it belongs with the metadata inside the netcdf file. What's the use case for storing it there? Because the problem remains that, unless you're storing and serving that dataset as a single 1 TB file that never gets modified or subset, as soon as anything at all happens to the file, those min and max values become tainted and unreliable, and ought to be recomputed. I could probably get behind it if it were called something different that highlighted that unreliability, like nominal_range, or display_range, or something like that, but calling it actual_range just seems to me like it's going to be misleading and incorrect dangerously often. Cheers, --Seth _______________________________________________ CF-metadata mailing list [email protected] http://mailman.cgd.ucar.edu/mailman/listinfo/cf-metadata
