Ben recently added a method to let the user control the output precision for ASCII Tecplot files, which is a great idea, but the default is 6 digits, which worries me. I vaguely remember struggling to hunt down some verification test failure years ago which turned out to be because the (GMV? XDA?) output truncated after 6 decimal places.
Could we default to 16 digits instead? Or perhaps set a typeof(Real) dependent DIGITS or PRECISION macro the way we do with TOLERANCE? I'd prefer precise output that an informed user has to override to get more efficiency over efficient output that an informed user has to override to get full precision. --- Roy ------------------------------------------------------------------------------ Download Intel® Parallel Studio Eval Try the new software tools for yourself. Speed compiling, find bugs proactively, and fine-tune applications for parallel performance. See why Intel Parallel Studio got high marks during beta. http://p.sf.net/sfu/intel-sw-dev _______________________________________________ Libmesh-devel mailing list [email protected] https://lists.sourceforge.net/lists/listinfo/libmesh-devel
