Just in case anybody else comes up against this one.  If someone is 
starting the search server from the same directory that you are using to 
run dfs and mapreduce as I was  then the search server picks ups the 
hadoop-site.xml and nutch-site.xml files that are in conf.  If those 
files point the filesystem to a DFS then the search server will be 
searching the DFS for the directory you provide upon startup.  If you 
are searching a local filesystem for the index you will need to point 
the filesystem to local in the configuration xml files.

My problem was that I was searching the DFS for a local path.  I thought 
that when I started the search server it took a local path, but it 
doesn't.  It takes a path and then finds which filesystem to use from 
the configuration xml files, in my case in the conf directory.  When I 
changed it to local, everything worked fine.

Dennis

Dennis Kubes wrote:
> Can someone explain how to setup distributed searching.  I have a 
> nutch-site.xml file setup like this:
>
> <configuration>
>
>  <property>
>    <name>fs.default.name</name>
>    <value>local</value>
>  </property>
>
>  <property>
>    <name>searcher.dir</name>
>    <value>C:\SERVERS\tomcat\webapps\ROOT\WEB-INF\classes</value>
>  </property>
>
> </configuration>
>
> With a search-servers.txt file in the classes directory.  The servers 
> file has entries that look like this:
>
> andromeda02 2388
>
> The server on the machine is started and running with this command.
>
> bin/nutch server 2388 /d01/crawl
>
> I can see it running and I can see the connect output, but searches 
> return no results.  The /d01/crawl directory contains indexes which is 
> a merged indexes directory, linkdb, crawldb, and segments.  Am I doing 
> something obviously wrong here?
>
> Dennis


_______________________________________________
Nutch-general mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/nutch-general

Reply via email to