Keep in mind that you do in fact lose the duplicates completely. I would think you want to keep one of the duplicates.

So for example if the merged files give you 4 records total:

John King
John King
Peter Rabbit
Bugs Bunny


You end up with 2 records - Peter Rabbit and Bugs Bunny - losing John King completely. The solution I presented returns all 3 unique records, losing just one of the repeated names instead of both, which to me makes more sense. Obviously you know your application better than I do, just want to make sure you're not surprised when 25% (or whatever percentage) of your database goes missing.


On 7/28/2012 2:22 PM, Lynne Stevens wrote:



*Hey that sounds good to me I will have Thomas bring over a set of the files
and I will use it . . will make Thomas jump up and down and happy


Defense

*Ok Thomas downloads 2 Data bases ( CVS format ) he wants to remove
Duplicates from Data Base B using Data base A and THEN Data Base A gets
deleted when done
*


OK, that is

$ cat fileA.csv fileB.csv | sort | uniq > fileC.csv








--
Andrew Brager
Green Gold Real Estate
Bakersfield, CA 93307
661 412 3304


--
For unsubscribe instructions e-mail to: [email protected]
Problems? http://www.libreoffice.org/get-help/mailing-lists/how-to-unsubscribe/
Posting guidelines + more: http://wiki.documentfoundation.org/Netiquette
List archive: http://listarchives.libreoffice.org/global/users/
All messages sent to this list will be publicly archived and cannot be deleted

Reply via email to