Thanks for posting your method of testing.
What I do on live data where running all the combos of HASH.AID isn't feasible (millions of records) is to take a random sample of the file and copy it into something manageable. Then using RESIZE and GROUP.STAT (since the records are usually very "lumpy"), I compare percent std deviations to look for record distribution.
--
Regards,
Clif
On Aug 14, 2004, at 12:16, Rosenberg Ben wrote:
Using a sample of files with no very large records,
or using id-only test files with null @RECORD,
for each filename, do
{
CLEAR-FILE DATA HASH.AID.FILE
for a sample of reasonable moduli, do
{
PHANTOM HASH.AID filename 2,18 mod sep
}
SORT HASH.AID.FILE BY-DSND LARGEST.GROUP
to see the worst file types.
}
------- u2-users mailing list [EMAIL PROTECTED] To unsubscribe please visit http://listserver.u2ug.org/
