Re: [GENERAL] Using database to find file doublettes in my computer

2008-11-20 Thread Craig Ringer
Lothar Behrens wrote: I have expected a smaller amount of records due to the fact that for 4 files each available 2 times (sum = 8) I have 8 records in ECADFiles, but must have 4 in the above result. So for an average of 2 doubles I expected half the files from ECADFiles, because one is exactly

Re: [GENERAL] Using database to find file doublettes in my computer

2008-11-19 Thread Lothar Behrens
On 18 Nov., 07:40, [EMAIL PROTECTED] (Craig Ringer) wrote: > -- Once paths is populated, extract duplicates: > > SELECT get_filename(path) AS fn, count(path) AS n > FROM paths HAVING count(path) > 1 > INTO TEMPORARY TABLE dup_files; > > -- Creates UNIQUE index on PATH as well > ALTER TABLE dup_fil

Re: [GENERAL] Using database to find file doublettes in my computer

2008-11-18 Thread Gerhard Heift
On Tue, Nov 18, 2008 at 12:36:42PM +, Sam Mason wrote: > On Mon, Nov 17, 2008 at 11:22:47AM -0800, Lothar Behrens wrote: > > I have a problem to find as fast as possible files that are double or > > in other words, identical. > > Also identifying those files that are not identical. > > I'd pro

Re: [GENERAL] Using database to find file doublettes in my computer

2008-11-18 Thread Sam Mason
On Mon, Nov 17, 2008 at 11:22:47AM -0800, Lothar Behrens wrote: > I have a problem to find as fast as possible files that are double or > in other words, identical. > Also identifying those files that are not identical. I'd probably just take a simple Unix command line approach, something like:

Re: [GENERAL] Using database to find file doublettes in my computer

2008-11-17 Thread Craig Ringer
Lothar Behrens wrote: > But how to query for files to display a 'left / right view' for each > file that is on multible places ? One approach is to use a query to extract the names of all files with duplicates, and store the results in a TEMPORARY table with a UNIQUE index (or PRIMARY KEY) on the

Re: [GENERAL] Using database to find file doublettes in my computer

2008-11-17 Thread Eus
Hi Ho! --- On Tue, 11/18/08, Lothar Behrens <[EMAIL PROTECTED]> wrote: > Hi, > > I have a problem to find as fast as possible files that are > double or > in other words, identical. > Also identifying those files that are not identical. > > My approach was to use dir /s and an awk script to con