Thanks for responding.
Byron Miller a �crit :
Actually at mozdex we have consolidated a bit and we are rebuilding under
the latest release. For 50 million urls a 200 gig disk is all you need.
That leaves you enough room for your segmetns, db and the space needed to
process (about double your db size)
Thanks, I think it's a googd starting point to make the test platform. I
will see after if I need to upgrade.
The biggest boost you can give your query servers is tons of memory. SATA
150 or Scsi drives at 10krpm is also a bonus.
I've seen that mozdex query servers use 2G of memory, do you think that
1G is enough for little trafic ?
For the categorizer, do you think it's possible, I took a look on some
Nutch plugins like GeoPosition and Language Identifier, but it doesn't
help me ?
If it's possible to do it, I will study the plugins deeper.