Just from a total noob standpoint (just installed my first LAMP box over the 
last month) realizing that I needed to look in the Runtime folder when I 
downloaded the tar.gz file was a HUGE step. 

Then we all run the Crawl at least to make sure things work.  The main tutorial 
was missing the [-solr] part of the crawl command line to get that to index.  
It wasn't after someone helped me here and pointed me to the actual documents 
that I found it.

Those were the 2 big things for me as a total noob, otherwise I'm really happy 
to have at least that part working.  Now, my stupid CentOS install only has 
libxml2 2.6.15 and I need 2.6.17 for php and I'm a few revisions off on libcurl 
also.  I have NO idea how to go back and fix that.  Not sure if I should just 
try to upgrade to php53 and hope for the best or what.  But, that's more of a 
solr / php question than a Nutch question I think.


-----Original Message-----
From: Markus Jelsma [mailto:[email protected]] 
Sent: Monday, July 11, 2011 3:19 PM
To: [email protected]
Cc: lewis john mcgibbney
Subject: Re: Nutch Gotchas as of release 1.3

Well, now i'm thinking of it: yes.

- there were three (incl. myself) people mentioning the problem described in 
NUTCH-1016;
- a few users don't seem to catch the part of the tutorial telling them to add 
their robot to the config
- missing crawl-urlfilter
- mails about missing solrUrl

I think quite a few users still rely on the crawl command instead of running a 
script.

> Hello list,
> 
> Do we have any suggestions we wish to discuss regarding the above?
> 
> thanks

Reply via email to