I have several *very* large files on which I need to perform some file sizing 
diagnostics.  Rather than repeatedly running HASH.AID against these files is 
there a good way to sample say 2-3 million records to copy into a test file?  
SAMPLE will only grab the first n records in hash order and I'm thinking that 
would not necessarily be a good representative sample of the file's contents.  
Am I up in the night thinking this is the case?  Is there a better way to get a 
good sample of records for this purpose?


Perry Taylor
Senior MV Architect
888 West Market Street, Suite 400
Louisville, KY 40202

CONFIDENTIALITY NOTICE: This e-mail message, including any 
attachments, is for the sole use of the intended recipient(s) 
and may contain confidential and privileged information.  Any
unauthorized review, use, disclosure or distribution is 
prohibited. ZirMed, Inc. has strict policies regarding the 
content of e-mail communications, specifically Protected Health 
Information, any communications containing such material will 
be returned to the originating party with such advisement 
noted. If you are not the intended recipient, please contact 
the sender by reply e-mail and destroy all copies of the 
original message.
U2-Users mailing list

Reply via email to