Andrzej Bialecki wrote:
 . How were the queries generated?  From a log or randomly?

Queries have been picked up manually, to test the worst performing cases from a real query log.

So, for example, the 50% error rate might not be typical, but could be worst-case.

 . When results differed greatly, did they look a lot worse?

Yes. E.g. see the differences for MAX_HITS=10000

The graph just shows that they differ, not how much better or worse they are, since the baseline is not perfect. When the top-10 is 50% different, are those 5 different hits markedly worse matches to your eye than the five they've displaced, or are they comparable? That's what really matters.

I actually forgot to write that I don't use any of Nutch code. Early on I decided to eliminate this part in order to get first the raw performance from Lucene - but still using the Lucene queries corresponding to translated Nutch queries.

What part of Nutch are you trying to avoid? Perhaps you could try measuring your Lucene-only benchmark against a Nutch-based one. If they don't differ markedly then you can simply use Nutch, which makes it a stronger benchmark. If they differ, then we should figure out why.

In several installations I use smaller values of slop (around 20-40). But this is motivated by better quality matches, not by performance, so I didn't test for this...

But that's a great reason to test for it! If lower slop can improve result quality, then we should certainly see if it also makes optimizations easier.

Doug

Reply via email to