Quanah Gibson-Mount wrote:
--On October 7, 2009 11:09:18 PM +0200 Emmanuel Lecharny
<[email protected]> wrote:
Quanah Gibson-Mount wrote:
--On October 7, 2009 3:32:51 PM -0400 Aaron Richton
<[email protected]> wrote:
On Wed, 7 Oct 2009, iz1ksw iz1ksw wrote:
What is the fastest way (in terms of openldap settings) to perform a
massive load (~200MB ldif file) of data into openldap directory?
Try "slapadd -q" (read slapadd(8) man page to get started).
slapadd -q is important, but so is having a large enough BDB cache in
the DB_CONFIG file, plus setting the right number of tool threads. I
took a load down from 17 hours to less than 2 hours adjusting all of
those correctly and adding -q.
The LDIF file is only 200Mb large. It should be a matter of minutes to
load it on a decent machine, even when using the standard parameters :)
There is not a 1 to 1 correlation between LDIF file size and resulting
database size. My LDIF file was 300MB and the resulting database was
12.5 GB.
Sure enough. My own test used a 200 Mb ldif file for a 6 Gb base.
I guess that it's all depending on the number of indexes. the best is to
start with a smaller file, evaluate the time it will take to load it,
and then try to extrapolate to the full file (roughly).
In any case, 200Mb for a ldif file will represent something between 100
000 to 1 000 000 entries, not a lot more (but it can be a much lower
number if you have JpegPhotos ...).
That being said, I doubt that the injection of a 200 Mb ldif will take
hours...
--
--
cordialement, regards,
Emmanuel Lécharny
www.iktek.com
directory.apache.org