Re: Real time indexing and distribution to lucene on separate boxes (long)

2004-03-12 Thread Otis Gospodnetic

--- Kevin A. Burton [EMAIL PROTECTED] wrote:
 Otis Gospodnetic wrote:
 
 I like option 3.  I've done it before, and it worked well.  I dealt
 with very small indices, though, and if your indices are several
 tens
 or hundred gigs, this may be hard for you.
 
 Option 4: search can be performed on an index that is being modified
 (update, delete, insert, optimize).  You'd just have to make sure
 not
 to recreate new IndexSearcher too frequently, if your index is being
 modified often.  Just change it every X index modification or every
 X
 minutes, and you'll be fine.
   
 
 Right now I'm thinking about #4... Disk may be cheap but a fast RAID
 10 
 array with 100G twice isn't THAT cheap... That's the worse case

Yes, but not everything needs to be on a fast RAID (you probabably are
using SCSI disks in RAID, which is what makes it expensive.  RAID
requires only a RAID controller).
You could have a Searcher machine with a set of cheap EIDE disks, and
use those as a copy target disks, which are not searched.
Once you transfer your indices there, you copy them on fast SCSI RAID
disks.

 Also... since the new indexes are SO small (~100M) the merges would 
 probably be easier on the machine than just doing a whole new write. 
 Of 
 course it's hard to make that argument with a 100G RAID array but
 we're 
 using rysnc to avoid distribution of network IO so the CPU
 computation 
 and network read would slow things down.
 
 The only way around this is the re-upload the whole 100G index but
 even 
 over gigabit ethernet this will take 15 minutes.  This doesn't scale
 as we add more searchers.

I wonder what happens if you try compressing the indices before copying
them over the network.
I wonder if it makes a difference whether you use compound vs.
traditional directories.
I wonder what the index size is if you use DbDirectory instead of
FSDirectory.

Otis


-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]



Real time indexing and distribution to lucene on separate boxes (long)

2004-03-11 Thread Kevin A. Burton
I'm curious to find out what others are doing in this situation.

I have two boxes... the indexer and the searcher.  The indexer is taking
documents and indexing them and creating indexes in a RAMDirectory (for
efficiency) and is then writing these indexes to disk as we begin to run 
out of
memory.  Usually these aren't very big... 15-100M or so.

Obviously I'm dividing the indexing and searching onto dedicated boxes to
improve efficiency.  The real issue though is that the searchers need to 
be live
all the time as indexes are being added at runtime.

So if that wasn't clear.  I actually have to push out fresh indexes 
WHILE users
are searching them.  Not a very easy thing to do.

Here's my question.  What are the optimum ways to then distribute these 
index
segments to the secondary searcher boxes.  I don't want to use the 
MultiSearcher
because it's slow once we have too many indexes (see my PS)

Here's what I'm currently thinking:

1.  Have the indexes sync'd to the searcher as shards directly.  This 
doesn't
scale as I would have to use the MultiSearcher which is slow when it has too
many indexes.  (And ideally we would want an optimized index).

2. Merge everything into one index on the indexer.  Lock the searcher, 
then copy
over the new index via rsync.  The problem here is that the searcher 
would need
to lock up while the sync is happening to prevent reads on the index.  
If I do
this enough and the system is optimzed I think I would only have to 
block for 5
seconds or so but that's STILL very long.

3. Have two directories on the searcher.  The indexer would then sync to 
a tmp
directory and then at run time swap them via a rename once the sync is over.
The downside here is that this will take up 2x disk space on the 
searcher.  The
upside is that the box will only slow down while the rsync is happening.

4. Do a LIVE index merge on the production box.  This might be an 
interesting
approach.  The major question I have is whether you can do an 
optimize/merge on
an index that's currently being used.  I *think* it might be possible 
but I'm
not sure.  This isn't as fast as performing the merge on the indexer 
before hand
but it does have the benefits of both worlds.

If anyone has any other ideas I would be all ears...

PS.. Random question.  The performance of the MultiSearcher is Mlog(N) 
correct?
Where N is the number of documents in the index and M is the number of 
indexes?
Is this about right?

--

Please reply using PGP.

   http://peerfear.org/pubkey.asc
   
   NewsMonster - http://www.newsmonster.org/
   
Kevin A. Burton, Location - San Francisco, CA, Cell - 415.595.9965
  AIM/YIM - sfburtonator,  Web - http://peerfear.org/
GPG fingerprint: 5FB2 F3E2 760E 70A8 6174 D393 E84D 8D04 99F1 4412
 IRC - freenode.net #infoanarchy | #p2p-hackers | #newsmonster



signature.asc
Description: OpenPGP digital signature


Re: Real time indexing and distribution to lucene on separate boxes (long)

2004-03-11 Thread Doug Cutting
Kevin A. Burton wrote:
  3. Have two directories on the searcher.  The indexer would then 
sync to
a tmp
directory and then at run time swap them via a rename once the sync is 
over.
The downside here is that this will take up 2x disk space on the 
searcher.  The
upside is that the box will only slow down while the rsync is happening.
For maximal search performance, this is your best bet.  Disk space is 
cheap.  At some point all newly issued queries start going against the 
new index, and, pretty soon, you can close and delete the old index. 
But you never have to stop searching.  Disk i/o will spike a bit during 
the changeover, as the new working set is swapped in.

Note that, if folks are paging through results by re-querying (the 
standard method) and the index is updated then things can get funky. 
One approach is to hang onto the old index longer, to make this less 
likely.  In any case, you might want to add an index-id to the search 
parameters, so that, if a next-page is issued when the index is no 
longer there you can give some sort of stale query error.

PS.. Random question.  The performance of the MultiSearcher is Mlog(N) 
correct?
Where N is the number of documents in the index and M is the number of 
indexes?
Is this about right?
No.

The added cost of a MultiSearcher is mostly proportional to just M, the 
number of indexes.  The normal cost of searching is still mostly 
proportional to N, the number of documents.  So M+N would probably be 
more accurate.  There is a log(M) here and there, to, e.g., figure out 
which index a doc id belongs to, but I doubt these are significant.

The significant costs of a MultiSearcher over an IndexSearcher are that 
it adds more term dictionary reads (one per query term per index) and 
more seeks (also one per query term per index, or two if you're using 
phrases).

Doug

-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]


Re: Real time indexing and distribution to lucene on separate boxes (long)

2004-03-11 Thread Kevin A. Burton
Otis Gospodnetic wrote:

I like option 3.  I've done it before, and it worked well.  I dealt
with very small indices, though, and if your indices are several tens
or hundred gigs, this may be hard for you.
Option 4: search can be performed on an index that is being modified
(update, delete, insert, optimize).  You'd just have to make sure not
to recreate new IndexSearcher too frequently, if your index is being
modified often.  Just change it every X index modification or every X
minutes, and you'll be fine.
 

Right now I'm thinking about #4... Disk may be cheap but a fast RAID 10 
array with 100G twice isn't THAT cheap... That's the worse case scenario 
of course and most modern search clusters use cheap hardware though...

Also... since the new indexes are SO small (~100M) the merges would 
probably be easier on the machine than just doing a whole new write.  Of 
course it's hard to make that argument with a 100G RAID array but we're 
using rysnc to avoid distribution of network IO so the CPU computation 
and network read would slow things down.

The only way around this is the re-upload the whole 100G index but even 
over gigabit ethernet this will take 15 minutes.  This doesn't scale as 
we add more searchers.

Thanks for the feedback!  I think now tha tI know that optmize is safe 
as long as I don't create a new reader... I'll be fine.  I do have think 
about how I'm going to handle search result nav.

Kevin

--

Please reply using PGP.

   http://peerfear.org/pubkey.asc
   
   NewsMonster - http://www.newsmonster.org/
   
Kevin A. Burton, Location - San Francisco, CA, Cell - 415.595.9965
  AIM/YIM - sfburtonator,  Web - http://peerfear.org/
GPG fingerprint: 5FB2 F3E2 760E 70A8 6174 D393 E84D 8D04 99F1 4412
 IRC - freenode.net #infoanarchy | #p2p-hackers | #newsmonster



signature.asc
Description: OpenPGP digital signature


Re: Real time indexing and distribution to lucene on separate boxes (long)

2004-03-11 Thread Dror Matalon
To clarify how option 3 works:

You have dira where the search is done and dirb where the indexing is
done. dirb grows when you add new items to it, and at some point you
swap and dirb becomes dira, but what do you do then?

Also, how do you write from the indexer to the directory on the search box?

We had two different ideas on how to do this, we ended up
implementing option 2, but you need to have support for snapshots to be
able to do it.

1. Similar idea to what you were suggesting, but optimizing the index:
1. The searcher reads from dira.
2. The writer writes to dirb which contains only *new* documents that
aren't in dira.
3. Periodically the writer stops writing and merges/optimizes dira and
dirb to dirc.
4. When done, the writer removes dirb and starts writing to it new
documents. 
5. The searcher notices that dirc is ready. Renames dira to something
else, renames dirc to dira.


The main problem with this approach is that the merging/optimizing can
be slow on large indexes. Even on fast machines with fast disks merging
several Gigs takes a while.

2. The index is NFS mounted. The indexer keeps writing to the index, and
at defined times, creates a NFS snapshot of the index. It then creates
an entry in a db to let the searcher know that a new snapshot has been
created.
The searcher checks once a minute the db to see if there's a new
snapshot. If there is one, it opens the index in the new snapshot and
swaps it for the old one. The code to do this is synchronized.

The nice thing about this solution is that you don't have just one copy
of the index and don't do any copying. But you need to use NFS and
snapshots.


Dror

On Thu, Mar 11, 2004 at 09:21:07AM -0800, Doug Cutting wrote:
 Kevin A. Burton wrote:
   3. Have two directories on the searcher.  The indexer would then 
 sync to
 a tmp
 directory and then at run time swap them via a rename once the sync is 
 over.
 The downside here is that this will take up 2x disk space on the 
 searcher.  The
 upside is that the box will only slow down while the rsync is happening.
 
 For maximal search performance, this is your best bet.  Disk space is 
 cheap.  At some point all newly issued queries start going against the 
 new index, and, pretty soon, you can close and delete the old index. 
 But you never have to stop searching.  Disk i/o will spike a bit during 
 the changeover, as the new working set is swapped in.
 
 Note that, if folks are paging through results by re-querying (the 
 standard method) and the index is updated then things can get funky. 
 One approach is to hang onto the old index longer, to make this less 
 likely.  In any case, you might want to add an index-id to the search 
 parameters, so that, if a next-page is issued when the index is no 
 longer there you can give some sort of stale query error.
 
 PS.. Random question.  The performance of the MultiSearcher is Mlog(N) 
 correct?
 Where N is the number of documents in the index and M is the number of 
 indexes?
 Is this about right?
 
 No.
 
 The added cost of a MultiSearcher is mostly proportional to just M, the 
 number of indexes.  The normal cost of searching is still mostly 
 proportional to N, the number of documents.  So M+N would probably be 
 more accurate.  There is a log(M) here and there, to, e.g., figure out 
 which index a doc id belongs to, but I doubt these are significant.
 
 The significant costs of a MultiSearcher over an IndexSearcher are that 
 it adds more term dictionary reads (one per query term per index) and 
 more seeks (also one per query term per index, or two if you're using 
 phrases).
 
 Doug
 
 -
 To unsubscribe, e-mail: [EMAIL PROTECTED]
 For additional commands, e-mail: [EMAIL PROTECTED]
 

-- 
Dror Matalon
Zapatec Inc 
1700 MLK Way
Berkeley, CA 94709
http://www.fastbuzz.com
http://www.zapatec.com

-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]