Re: [U2] [UV] Record Sampling for File Sizing Diagnostics

2013-06-12 Thread Oaks, Harold
Perry:
In Universe the modifier SAMPLED nn selects every nn'th record.  For
example, SELECT MYFILE SAMPLED 50 selects every 50th record.  That way
you plow thru the entire file but come back with only  1/50 of the
records.   I don't know if this keyword exists in Unidata
Harold Oaks
Clark County, WA

-Original Message-
From: u2-users-boun...@listserver.u2ug.org
[mailto:u2-users-boun...@listserver.u2ug.org] On Behalf Of Perry Taylor
Sent: Tuesday, June 11, 2013 9:01 AM
To: U2-Users List
Subject: [U2] [UV] Record Sampling for File Sizing Diagnostics

I have several *very* large files on which I need to perform some file
sizing diagnostics.  Rather than repeatedly running HASH.AID against
these files is there a good way to sample say 2-3 million records to
copy into a test file?  SAMPLE will only grab the first n records in
hash order and I'm thinking that would not necessarily be a good
representative sample of the file's contents.  Am I up in the night
thinking this is the case?  Is there a better way to get a good sample
of records for this purpose?

Thanks.
Perry

Perry Taylor
Senior MV Architect
ZirMed
888 West Market Street, Suite 400
Louisville, KY 40202
www.zirmed.comhttp://www.zirmed.com/



CONFIDENTIALITY NOTICE: This e-mail message, including any attachments,
is for the sole use of the intended recipient(s) and may contain
confidential and privileged information.  Any unauthorized review, use,
disclosure or distribution is prohibited. ZirMed, Inc. has strict
policies regarding the content of e-mail communications, specifically
Protected Health Information, any communications containing such
material will be returned to the originating party with such advisement
noted. If you are not the intended recipient, please contact the sender
by reply e-mail and destroy all copies of the original message.
___
U2-Users mailing list
U2-Users@listserver.u2ug.org
http://listserver.u2ug.org/mailman/listinfo/u2-users

This e-mail and related attachments and any response may be subject to public 
disclosure under state law.
___
U2-Users mailing list
U2-Users@listserver.u2ug.org
http://listserver.u2ug.org/mailman/listinfo/u2-users


Re: [U2] [UV] Record Sampling for File Sizing Diagnostics

2013-06-12 Thread Bill Haskett

Harold:

UniData does have the SAMPLED verb.  It works the same way as described.

Bill


- Original Message -
*From:* harold.o...@clark.wa.gov
*To:* U2 Users List u2-users@listserver.u2ug.org
*Date:* 6/12/2013 2:17 PM
*Subject:* Re: [U2] [UV] Record Sampling for File Sizing Diagnostics

Perry:
In Universe the modifier SAMPLED nn selects every nn'th record.  For
example, SELECT MYFILE SAMPLED 50 selects every 50th record.  That way
you plow thru the entire file but come back with only  1/50 of the
records.   I don't know if this keyword exists in Unidata
Harold Oaks
Clark County, WA

-Original Message-
From: u2-users-boun...@listserver.u2ug.org
[mailto:u2-users-boun...@listserver.u2ug.org] On Behalf Of Perry Taylor
Sent: Tuesday, June 11, 2013 9:01 AM
To: U2-Users List
Subject: [U2] [UV] Record Sampling for File Sizing Diagnostics

I have several *very* large files on which I need to perform some file
sizing diagnostics.  Rather than repeatedly running HASH.AID against
these files is there a good way to sample say 2-3 million records to
copy into a test file?  SAMPLE will only grab the first n records in
hash order and I'm thinking that would not necessarily be a good
representative sample of the file's contents.  Am I up in the night
thinking this is the case?  Is there a better way to get a good sample
of records for this purpose?

Thanks.
Perry

Perry Taylor
Senior MV Architect
ZirMed
888 West Market Street, Suite 400
Louisville, KY 40202
www.zirmed.comhttp://www.zirmed.com/



CONFIDENTIALITY NOTICE: This e-mail message, including any attachments,
is for the sole use of the intended recipient(s) and may contain
confidential and privileged information.  Any unauthorized review, use,
disclosure or distribution is prohibited. ZirMed, Inc. has strict
policies regarding the content of e-mail communications, specifically
Protected Health Information, any communications containing such
material will be returned to the originating party with such advisement
noted. If you are not the intended recipient, please contact the sender
by reply e-mail and destroy all copies of the original message.
___
U2-Users mailing list
U2-Users@listserver.u2ug.org
http://listserver.u2ug.org/mailman/listinfo/u2-users

This e-mail and related attachments and any response may be subject to public 
disclosure under state law.
___
U2-Users mailing list
U2-Users@listserver.u2ug.org
http://listserver.u2ug.org/mailman/listinfo/u2-users


___
U2-Users mailing list
U2-Users@listserver.u2ug.org
http://listserver.u2ug.org/mailman/listinfo/u2-users


Re: [U2] [UV] Record Sampling for File Sizing Diagnostics

2013-06-11 Thread Charles Stevenson
SAMPLED keyword with a D on the end will go thru the entire file.
SAMPLED 1000  will get every thousandth record as it reads through the file.
SAMPLED 1000 SAMPLE 2000  will do the same, but then stop after it builds a
list of 2000. That is, after it's read thru the 1st 2 million keys.


On Tue, Jun 11, 2013 at 11:01 AM, Perry Taylor perry.tay...@zirmed.comwrote:

 I have several *very* large files on which I need to perform some file
 sizing diagnostics.  Rather than repeatedly running HASH.AID against these
 files is there a good way to sample say 2-3 million records to copy into a
 test file?  SAMPLE will only grab the first n records in hash order and I'm
 thinking that would not necessarily be a good representative sample of the
 file's contents.  Am I up in the night thinking this is the case?  Is there
 a better way to get a good sample of records for this purpose?

 Thanks.
 Perry

 Perry Taylor
 Senior MV Architect
 ZirMed
 888 West Market Street, Suite 400
 Louisville, KY 40202
 www.zirmed.comhttp://www.zirmed.com/



 CONFIDENTIALITY NOTICE: This e-mail message, including any
 attachments, is for the sole use of the intended recipient(s)
 and may contain confidential and privileged information.  Any
 unauthorized review, use, disclosure or distribution is
 prohibited. ZirMed, Inc. has strict policies regarding the
 content of e-mail communications, specifically Protected Health
 Information, any communications containing such material will
 be returned to the originating party with such advisement
 noted. If you are not the intended recipient, please contact
 the sender by reply e-mail and destroy all copies of the
 original message.
 ___
 U2-Users mailing list
 U2-Users@listserver.u2ug.org
 http://listserver.u2ug.org/mailman/listinfo/u2-users

___
U2-Users mailing list
U2-Users@listserver.u2ug.org
http://listserver.u2ug.org/mailman/listinfo/u2-users


Re: [U2] [UV] Record Sampling for File Sizing Diagnostics

2013-06-11 Thread Charlie Noah

Hi Perry,

Does UV support SAMPLED? If so, use:

SELECT file SAMPLE 300 SAMPLED 

where  is the number of records to skip between sample IDs picked 
up. If it doesn't, you're probably going to have to write a little program.


Regards,
Charlie

Tiny Bear's Wild Bird Store
Home of Safety Net Shipping
http://www.TinyBearWildBirdStore.com
Toll Free: 1-855-TinyBear (855-846-9232)

On 06-11-2013 11:01 AM, Perry Taylor wrote:

I have several *very* large files on which I need to perform some file sizing 
diagnostics.  Rather than repeatedly running HASH.AID against these files is 
there a good way to sample say 2-3 million records to copy into a test file?  
SAMPLE will only grab the first n records in hash order and I'm thinking that 
would not necessarily be a good representative sample of the file's contents.  
Am I up in the night thinking this is the case?  Is there a better way to get a 
good sample of records for this purpose?

Thanks.
Perry

Perry Taylor
Senior MV Architect
ZirMed
888 West Market Street, Suite 400
Louisville, KY 40202
www.zirmed.comhttp://www.zirmed.com/



CONFIDENTIALITY NOTICE: This e-mail message, including any
attachments, is for the sole use of the intended recipient(s)
and may contain confidential and privileged information.  Any
unauthorized review, use, disclosure or distribution is
prohibited. ZirMed, Inc. has strict policies regarding the
content of e-mail communications, specifically Protected Health
Information, any communications containing such material will
be returned to the originating party with such advisement
noted. If you are not the intended recipient, please contact
the sender by reply e-mail and destroy all copies of the
original message.
___
U2-Users mailing list
U2-Users@listserver.u2ug.org
http://listserver.u2ug.org/mailman/listinfo/u2-users



___
U2-Users mailing list
U2-Users@listserver.u2ug.org
http://listserver.u2ug.org/mailman/listinfo/u2-users


Re: [U2] [UV] Record Sampling for File Sizing Diagnostics

2013-06-11 Thread Charlie Noah

I probably hit Send within milliseconds of you. :-)
Charlie

On 06-11-2013 11:15 AM, Charles Stevenson wrote:

SAMPLED keyword with a D on the end will go thru the entire file.
SAMPLED 1000  will get every thousandth record as it reads through the file.
SAMPLED 1000 SAMPLE 2000  will do the same, but then stop after it builds a
list of 2000. That is, after it's read thru the 1st 2 million keys.


On Tue, Jun 11, 2013 at 11:01 AM, Perry Taylor perry.tay...@zirmed.comwrote:


I have several *very* large files on which I need to perform some file
sizing diagnostics.  Rather than repeatedly running HASH.AID against these
files is there a good way to sample say 2-3 million records to copy into a
test file?  SAMPLE will only grab the first n records in hash order and I'm
thinking that would not necessarily be a good representative sample of the
file's contents.  Am I up in the night thinking this is the case?  Is there
a better way to get a good sample of records for this purpose?

Thanks.
Perry

Perry Taylor
Senior MV Architect
ZirMed
888 West Market Street, Suite 400
Louisville, KY 40202
www.zirmed.comhttp://www.zirmed.com/



CONFIDENTIALITY NOTICE: This e-mail message, including any
attachments, is for the sole use of the intended recipient(s)
and may contain confidential and privileged information.  Any
unauthorized review, use, disclosure or distribution is
prohibited. ZirMed, Inc. has strict policies regarding the
content of e-mail communications, specifically Protected Health
Information, any communications containing such material will
be returned to the originating party with such advisement
noted. If you are not the intended recipient, please contact
the sender by reply e-mail and destroy all copies of the
original message.
___
U2-Users mailing list
U2-Users@listserver.u2ug.org
http://listserver.u2ug.org/mailman/listinfo/u2-users


___
U2-Users mailing list
U2-Users@listserver.u2ug.org
http://listserver.u2ug.org/mailman/listinfo/u2-users



___
U2-Users mailing list
U2-Users@listserver.u2ug.org
http://listserver.u2ug.org/mailman/listinfo/u2-users


Re: [U2] [UV] Record Sampling for File Sizing Diagnostics

2013-06-11 Thread Perry Taylor
Great minds!

Wow I knew about SAMPLED but it didn't occur to me that it could be used 
concurrently with SAMPLE.  Nice!  I'll give that try.

Thanks guys!

Perry

-Original Message-
From: u2-users-boun...@listserver.u2ug.org 
[mailto:u2-users-boun...@listserver.u2ug.org] On Behalf Of Charlie Noah
Sent: Tuesday, June 11, 2013 10:21 AM
To: U2 Users List
Subject: Re: [U2] [UV] Record Sampling for File Sizing Diagnostics

I probably hit Send within milliseconds of you. :-)
Charlie

On 06-11-2013 11:15 AM, Charles Stevenson wrote:
 SAMPLED keyword with a D on the end will go thru the entire file.
 SAMPLED 1000  will get every thousandth record as it reads through the file.
 SAMPLED 1000 SAMPLE 2000  will do the same, but then stop after it builds a
 list of 2000. That is, after it's read thru the 1st 2 million keys.


 On Tue, Jun 11, 2013 at 11:01 AM, Perry Taylor perry.tay...@zirmed.comwrote:

 I have several *very* large files on which I need to perform some file
 sizing diagnostics.  Rather than repeatedly running HASH.AID against these
 files is there a good way to sample say 2-3 million records to copy into a
 test file?  SAMPLE will only grab the first n records in hash order and I'm
 thinking that would not necessarily be a good representative sample of the
 file's contents.  Am I up in the night thinking this is the case?  Is there
 a better way to get a good sample of records for this purpose?

 Thanks.
 Perry

 Perry Taylor
 Senior MV Architect
 ZirMed
 888 West Market Street, Suite 400
 Louisville, KY 40202
 www.zirmed.comhttp://www.zirmed.com/



 CONFIDENTIALITY NOTICE: This e-mail message, including any
 attachments, is for the sole use of the intended recipient(s)
 and may contain confidential and privileged information.  Any
 unauthorized review, use, disclosure or distribution is
 prohibited. ZirMed, Inc. has strict policies regarding the
 content of e-mail communications, specifically Protected Health
 Information, any communications containing such material will
 be returned to the originating party with such advisement
 noted. If you are not the intended recipient, please contact
 the sender by reply e-mail and destroy all copies of the
 original message.
 ___
 U2-Users mailing list
 U2-Users@listserver.u2ug.org
 http://listserver.u2ug.org/mailman/listinfo/u2-users

 ___
 U2-Users mailing list
 U2-Users@listserver.u2ug.org
 http://listserver.u2ug.org/mailman/listinfo/u2-users


___
U2-Users mailing list
U2-Users@listserver.u2ug.org
http://listserver.u2ug.org/mailman/listinfo/u2-users
___
U2-Users mailing list
U2-Users@listserver.u2ug.org
http://listserver.u2ug.org/mailman/listinfo/u2-users