Hi, 

Our process isn't very finessed, but right now it's:

-run brunnhilde on the files to get reports, in this case we want the 
duplicates.csv report

-if requested, the archivist does manual review of that csv to decide which 
copy of the file should be retained. The copy that should be retained should be 
the first one in the list (i.e. if there are 4 copies of a file listed in the 
duplicates.csv, the first one will be kept). I can see though how that would be 
cumbersome if you had hundreds to correct/move. It could likely be scripted if 
they all followed a similar pattern (e.g. were all in the same path). 

-we run a little bash script that reads through that csv line by line. It looks 
at the checksum, if the checksum does not match the previous checksum in the 
list, it continues. If it does match the previous checksum in the list, then it 
moves that file to a Duplicates directory and logs that move, e.g. "MOVED 
<originalFilePath> to <directory>"

And, like others have said, none of this captures close duplicates.

Reply via email to