hpux, unidata 7.1.8
More in my quest to understand dynamic files in Unidata. I understand
that split and merge load are not universal constants and many file
factors affect the optimal values for them. What I don't understand is
why you'd want a split load of something like 20 in any circumstance.
Perhaps it's my misunderstanding of split load, but I take it to mean that
when a group reaches 20% full, it will split it into two different groups.
Of course, I'm dealing with a headache of a dynamic file that just isn't
happy. In short, it's got about 33,000,000 records, each in under 512 in
size with purely integer keys. I can create a dynamic file with hash type
0, modulo around 3,000,000 and block size 1024 which looks good at the
start: it's got 3 dat segments and 1 small ovr segment.
When I copy records from the original to the new dynamic file and get
somewhere around 20,000,000 records, it start creating ovr segments lie
crazy. I end up with 4 dat segments and 5 ovr segments.
When I run guide on resulting file, it's taking up about 9Gig on disk, but
only contains about 1.1Gig of data. I've tried split/merge loads of 20/10
and 60/40 with similar results.
Hence, I'm trying to determine why it's splitting like mad even thought
it's mostly empty. With the purely integer (sequential) keys, it's got a
nice even distribution in keys. I can avoid the overflow by creating
with a rediculously large modulo (eg: 10,000,000) but again, I get a file
that's 80% empty in the end.
Anyadvice appreciated.
Jeff Butera, Ph.D.
Administrative Systems
Hampshire College
[EMAIL PROTECTED]
413-559-5556
"Daddy - did you lose your mind?"
Catherine Butera
-------
u2-users mailing list
[email protected]
To unsubscribe please visit http://listserver.u2ug.org/