Dear Briggs, Currently we allocated 1gig for JVM and Resin/Tomcat.
Greetings, Joost Briggs wrote: > > How much memory are you currently allocating to the search servers? > > > > On 4/23/07, JoostRuiter <[EMAIL PROTECTED]> wrote: >> >> Hi All, >> >> First off, I'm quite the noob when it comes to Nutch, so don't bash me if >> the following is an enormously stupid question. >> >> We're using Nutch on a P4 Duo Core system (800mhz fsb) with 4gig RAM and >> a >> 500gig SATA (3gig/sec) HD. We indexed 350 000 pages into 1 segment of >> 15gig. >> >> >> Performance is really poor, if we do get search results it will take >> multiple minutes. When the query is longer we are getting the following: >> >> "java.lang.OutOfMemoryError: Java heap memory" >> >> What we have tried to improve on this: >> - Slice the segments into smaller chuncks (max: 50000 url/per seg) >> - Set io.map.index.skip to 8 >> - Set indexer.termIndexInterval to 1024 >> - Cluster with Hadoop (4 nodes to search) >> >> Any ideas? Missing information? Please let me know, this is my graduation >> internship and I would really like to get a good grade ;) >> -- >> View this message in context: >> http://www.nabble.com/Perfomance-problems-and-segmenting-tf3631982.html#a10141310 >> Sent from the Nutch - Dev mailing list archive at Nabble.com. >> >> > > > -- > "Conscious decisions by concious minds are what make reality real" > > -- View this message in context: http://www.nabble.com/Perfomance-problems-and-segmenting-tf3631982.html#a10141752 Sent from the Nutch - Dev mailing list archive at Nabble.com. ------------------------------------------------------------------------- This SF.net email is sponsored by DB2 Express Download DB2 Express C - the FREE version of DB2 express and take control of your XML. No limits. Just data. Click to get it now. http://sourceforge.net/powerbar/db2/ _______________________________________________ Nutch-developers mailing list Nutch-developers@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/nutch-developers