Understand why NRT performance is affected by flush frequency
-------------------------------------------------------------

                 Key: LUCENE-2143
                 URL: https://issues.apache.org/jira/browse/LUCENE-2143
             Project: Lucene - Java
          Issue Type: Bug
          Components: Index
            Reporter: Michael McCandless
            Assignee: Michael McCandless
             Fix For: 3.1


In LUCENE-2061 (perf tests for NRT), I test NRT performance by first
getting a baseline QPS with only searching, using enough threads to
saturate.

Then, I pick an indexing rate (I used 100 docs/sec), and index docs at
that rate, and I also reopen a NRT reader at different frequencies
(10/sec, 1/sec, every 5 seconds, etc.), and then again test QPS
(saturated).

I think this is a good approach for testing NRT -- apps can see, as a
function of "freshness" and at a fixed indexing rate, what the cost is
to QPS.  You'd expect as index rate goes up, and freshness goes up,
QPS will go down.

But I found something very strange: the low frequency reopen rates
often caused a highish hit to QPS.  When I forced IW to flush every
100 docs (= once per second), the performance was generally much
better.

I actually would've expected the reverse -- flushing in batch ought to
use fewer resoruces.

One theory is something odd about my test env (based on OpenSolaris),
so I'd like to retest on a more mainstream OS.

I'm opening this issue to get to the bottom of it...


-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


---------------------------------------------------------------------
To unsubscribe, e-mail: java-dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: java-dev-h...@lucene.apache.org

Reply via email to