Hi, I have an LDAP directory with a very large number of records (some possibly duplicated in their entirety or partially as superset/subset) which I would like to condense and repair and correct insofar as the individual subrecords/fields within each record are concerned.
The format of this LDAP directory, when dumped, is a set of millions of rows of data which when sorted and uniqued on the cname results in a small fraction of the original total (.00746% to be exact), though whether the duplicates themselves have same fields is another matter entirely. The records are in XML format and consist of key/value pairs. My suspicion is that this directory has never been properly maintained so I have some questions: + what are the accepted ways via automation to maintain this directory + what methods or code exist to condense and verify a hitherto-unmaintained LDAP directory + simplifying the data to the bare bones number of records, discarding the others after making a general full backup So what I am asking for is a general set of already written Perl tools using Net::LDAP which deal with LDAP directories intelligently. --Stuart