Anyone ?
Any help to understand this package is appreciated.

Thanks,
T

On Thu, May 28, 2009 at 3:18 PM, Tenaali Ram <tenaali...@gmail.com> wrote:

> Hi,
>
> I am trying to understand the code of index package to build a distributed
> Lucene index. I have some very basic questions and would really appreciate
> if someone can help me understand this code-
>
> 1) If I already have Lucene index (divided into shards), should I upload
> these indexes into HDFS and provide its location or the code will pick these
> shards from local file system ?
>
> 2) How is the code adding a document in the lucene index, I can see there
> is a index selection policy. Assuming round robin policy is chosen, how is
> the code adding a document in the lucene index? This is related to first
> question - is the index where the new document is to be added in HDFS or in
> local file system. I read in the README that the index is first created on
> local file system, then copied back to HDFS. Can someone please point me to
> the code that is doing this.
>
> 3) After the map reduce job finishes, where are the final indexes ? In HDFS
> ?
>
> 4) Correct me if I am wrong- the code builds multiple indexes, where each
> index is an instance of Lucene Index having a disjoint subset of documents
> from the corpus. So, if I have to search a term, I have to search each index
> and then merge the result. If this is correct, then how is the IDF of a term
> which is a global statistic computed and updated in each index ? I mean each
> index can compute the IDF wrt. to the subset of documents that it has, but
> can not compute the global IDF of a term (since it knows nothing about other
> indexes, which might have the same term in other documents).
>
> Thanks,
> -T
>
>
>

Reply via email to