Hi,
On 8/1/07, Dennis Kubes [EMAIL PROTECTED] wrote:
I am currently writing a python script to automate this whole process
from inject to pushing out to search servers. It should be done in a
day or two and I will post it on the wiki.
(it is a bit of shameless self-promotion but here it
Actually no. Let's say you have 10 machines and hence 10 search
servers. You would run through 10 iterations of fetch-index-deploy, one
to each machine. Lets say you have 3 million pages per machine so this
whole system could support a 30 million page index.
Once you deploy to 10 you would
Ah, OK, I get it. Sadly for me, this precise approach is probably not going
meet my requirements, but it really helps to get me going, and I think a
variation on it will suit me quite well. I'm very much looking forward to
seeing the script that automates this.
I have one minor quibble with
Thanks very much for the extended reply; lots of food for thought.
WRT the merge/index time on a large index, I kind of suspected this might be
the case. It's already taking a bit of time (albeit on a weak box) with my
relatively small index. In general the approach you outline sounds like
I am currently writing a python script to automate this whole process
from inject to pushing out to search servers. It should be done in a
day or two and I will post it on the wiki.
Dennis Kubes
charlie w wrote:
Thanks very much for the extended reply; lots of food for thought.
WRT the
On 8/1/07, Dennis Kubes [EMAIL PROTECTED] wrote:
I am currently writing a python script to automate this whole process
from inject to pushing out to search servers. It should be done in a
day or two and I will post it on the wiki.
I'm very much looking forward to this. Reading the code
It is not a problem to contact me directly if you have questions. I am
going to include this post on the mailing list as well in case other
people have similar questions.
When we originally started (and back when I wrote the tutorial), I
thought the best approache would be to have a single