Certainly others do agree with you to some degree that this case is on the cost/benefit borderline. Again, this case wasn't really the point.

My point was it feels to me that you have, on occasion, been over- quick to criticize without paying sufficiently respectful attention to the details of what is being discussed. For instance, the criticism of "these tests should be done on a *nix platform" to someone who has repeated the tests on osx (yes, a nix) and windows. Or that the test is too short and the index in memory (it was 10MM docs with term vecs on FSDirectory. It is possible that some of the index wasn't fsync'd at the end of each test, I suppose, but I would expect this to be a small amount and equivalent in pre- and post- patch scenarios). Or calling a full index run of 10MM docs a "micro benchmark".

I do think that I was unchill in sending the original post to the list instead of to you via personal mail. I shouldn't have.

regards,
-Mike

On 10-Feb-08, at 7:33 PM, robert engels wrote:

Please chill. You are inferring something that was not implied. You may think it lacks perspective and respect (I disagree on both), but it certainly doesn't lack in correctness.

First, depending on how you measure it, 2x speedup equates to a 50% reduction in time. In my review of the changes that brought about the biggest performance gains from 1.9 on, almost all were related to avoiding disk accesses by buffering more documents and doing more processing in memory. I don't think many of the micro- benchmarks mattered much, and with a JVM environment it is very difficult to prove as it is going to be heavily JVM and configuration dependent.

The main point was that ANY disk access is going to be ORDERS OF MAGNITUDE slower than any of these sort of optimizations.

So either you are loading the index completely in memory (only small indexes, so the difference in speed is not going to matter much), or you might be using a federated system of memory indices (to form a large index), but USUALLY at some point the index must be first created in a persistent store (that which is covered here), in order to provide realistic restart times, etc.

The author of the patch and timings gives no information as to disk speed, IO speed, controllers, raid configuration , etc. When creating an index in persistent store, these factors matter more than a 2-4% speed up. Creating an index completely in memory is then bound by the reading of the data from the disk, and/or the network - all much slower than the actual indexing.

Usually optimizations like this only matter in areas of development where the data set is small, but the processing large (a lot of numerical analysis). In some cases the data set may also be "large", but then usually the processing is exponentially larger. The building of the index in Lucene in not very computationally expensive.

If you are going spend hundreds of hours "optimizing", you best be optimizing the right things. That was the point of the link I sent (the quotes are from people far more capable than I).

I was trying to make the point that a 2-4 % speed up probably doesn't amount to much in a real environment given all of the other factors, so it is probably better for the project/community to err on the side of code clarity and ease of maintenance.

The project can continue to do what it wants (obviously) - but what I was pointing out should be nothing new to experienced designers/ developers - I only offering a reminder. It is my observation (others will disagree !), but I think a lot of Lucene has some unneeded esoteric code, where the benefit doesn't match the cost.

On Feb 10, 2008, at 5:48 PM, Mike Klaas wrote:

While I agree in general that excessive optimization at the expense of code clarity is undesirable, you are overstating the point. 2X is a ridiculous threshold to apply to something as performance critical as a full text search engine. If search was twice as slow, lucene would be utterly unusable for me. Indexing less important than search, of course, but a 2X slowdown with be quite painful there.

I don't have an opinion in this case: I believe that there is a tradeoff but that it is the responsibility of the commiter(s) to achieve the correct balance--they are the ones who will be maintaining the code, after all. I find your persistence surprising and your tone dangerously near condescending. Telling the guy who has spent hundreds of hours carefully optimizing this code that "Almost always there is a better bottleneck somewhere" shows an astonishing lack of perspective and respect.

-Mike

On 10-Feb-08, at 12:15 PM, robert engels wrote:

I am not sure these numbers matter. I think they are skewed because you are probably running too short a test, and the index is in memory (or OS cache).

Once you use a real index that needs to read/write from the disk, the percentage change will be negligible.

This is the problem with many of these "performance changes" - they just aren't real world enough. Even if they were, I would argue that code simplicity/maintainability is worth more than 6 seconds on a operation that takes 4 minutes to run...

There are many people that believe micro benchmarks are next to worthless. A good rule of thumb is that if the optimization doesn't result in 2x speedup, it probably shouldn't be done. In most cases any efficiency gains are later lost in maintainability issues.

See http://en.wikipedia.org/wiki/Optimization_(computer_science)

Almost always there is a better bottleneck somewhere.

On Feb 10, 2008, at 1:37 PM, Michael McCandless wrote:


Yonik Seeley wrote:

I wonder how well a single generic quickSort(Object[] arr, int low, int high) would perform vs the type-specific ones? I guess the main overhead would be a cast from Object to the specific class to do the
compare?  Too bad Java doesn't have true generics/templates.


OK I tested this.

Starting from the patch on LUCENE-1172, which has 3 quickSort methods (one per type), I created a single quickSort method on Object[] that
takes a Comparator, and made 3 Comparators instead.

Mac OS X 10.4 (JVM 1.5):

    original patch --> 247.1
  simplified patch --> 254.9 (3.2% slower)

Windows Server 2003 R64 (JVM 1.6):

    original patch --> 440.6
  simplified patch --> 452.7 (2.7% slower)

The times are best in 10 runs. I'm running all tests with these JVM
args:

  -Xms1024M -Xmx1024M -Xbatch -server

I think this is a big enough difference in performance that it's
worth keeping 3 separate quickSorts in DocumentsWriter.

Mike

------------------------------------------------------------------- --
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]



-------------------------------------------------------------------- -
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]



---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]



---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]



---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to