Thanks Uwe !

Le 9 janv. 2018 11:55 AM, "Uwe Schindler" <u...@thetaphi.de> a écrit :

I enabled the Policeman and ASF Jenkins builds of 7.2 branch.



Uwe



-----

Uwe Schindler

Achterdiek 19, D-28357 Bremen
<https://maps.google.com/?q=Achterdiek+19,+D-28357+Bremen&entry=gmail&source=g>

http://www.thetaphi.de

eMail: u...@thetaphi.de



*From:* jim ferenczi [mailto:jim.feren...@gmail.com]
*Sent:* Tuesday, January 9, 2018 11:00 AM
*To:* dev@lucene.apache.org
*Subject:* Re: BugFix release 7.2.1



Hi,

All the issues we discussed have been backported to 6.2. I added the draft
for the release notes in the wiki:

https://wiki.apache.org/lucene-java/ReleaseNote721

https://wiki.apache.org/solr/ReleaseNote721



I'll create the first RC later today.







2018-01-09 4:56 GMT+01:00 S G <sg.online.em...@gmail.com>:

Sorry, I missed some of the details but this is what we did in one of my
past projects with success:



We can begin by supporting only those machines where Apache Solr's
regression tests are run.

The aim is to identify OS-independent performance regressions, not to
certify each OS where Solr could be run.



Repository wise is easy too - We store the results in a performance-results
directory that stays in the github repo of Apache Solr.

This directory will receive metric-result-file(s) whenever a Solr release
is made.

And if older files are present, then the last metric file will be used to
compare the current performance.

When not making a release, the directory can be used to compare current
code's performance without writing to the performance-results directory.

When releasing Solr, the performance-metrics file should get updated
automatically.



Further improvements can include:

1) Deleting older files from performance-results directory

2) Having performance-results directories for each OS where Solr is
released (if we think OS-dependent performance issues could be there).



These ideas can be fine-tuned to ensure that they work.

Please suggest more issues if you think this would be impractical.



Thanks

SG









On Mon, Jan 8, 2018 at 12:59 PM, Erick Erickson <erickerick...@gmail.com>
wrote:

Hmmm, I think you missed my implied point. How are these metrics collected
and compared? There are about a dozen different machines running various op
systems etc. For these measurements to spot regressions and/or
improvements, they need to have a repository where the results get
published. So a report like "build XXX took YYY seconds to index ZZZ
documents" doesn't tell us anything. You need to gather then for a
_specific_ machine.

As for whether they should be run or not, an annotation could help here,
there are already @Slow, @Nightly, @Weekly and @Performance could be added.
Mike McCandless has some of these kinds of things already for Lucene, I
htink the first thing would be to check whether they are already done, it's
possible you'd be reinventing the wheel.

Best,

Erick



On Mon, Jan 8, 2018 at 11:45 AM, S G <sg.online.em...@gmail.com> wrote:

We can put some lower limits on CPU and Memory for running a performance
test.

If those lower limits are not met, then the test will just skip execution.



And then we put some lower bounds (time-wise) on the time spent by
different parts of the test like:

 - Max time taken to index 1 million documents

 - Max time taken to query, facet, pivot etc

 - Max time taken to delete 100,000 documents while read and writes are
happening.



For all of the above, we can publish metrics like 5minRate, 95thPercent and
assert on values lower than a particular value.



I know some other software compare CPU cycles across different runs as well
but not sure how.



Such tests will give us more confidence when releasing/adopting new
features like pint compared to tint etc.



Thanks

SG







On Sat, Jan 6, 2018 at 9:59 AM, Erick Erickson <erickerick...@gmail.com>
wrote:

Not sure how performance tests in the unit tests would be interpreted. If I
run the same suite on two different machines how do I compare the numbers?

Or are you thinking of having some tests so someone can check out different
versions of Solr and run the perf tests on a single machine, perhaps using
bisect to pinpoint when something changed?

I'm not opposed at all, just trying to understand how one would go about
using such tests.

Best,

Erick



On Fri, Jan 5, 2018 at 10:09 PM, S G <sg.online.em...@gmail.com> wrote:

Just curious to know, does the test suite include some performance test
also?

I would like to know the performance impact of using pints vs tints or ints
etc.

If they are not there, I can try to add some tests for the same.



Thanks

SG





On Fri, Jan 5, 2018 at 5:47 PM, Đạt Cao Mạnh <caomanhdat...@gmail.com>
wrote:

Hi all,



I will work on SOLR-11771
<https://issues.apache.org/jira/browse/SOLR-11771> today,
It is a simple fix and will be great if it get fixed in 7.2.1



On Fri, Jan 5, 2018 at 11:23 PM Erick Erickson <erickerick...@gmail.com>
wrote:

Neither of those Solr fixes are earth shatteringly important, they've both
been around for quite a while. I don't think it's urgent to include them.

That said, they're pretty simple and isolated so worth doing if Jim is
willing. But not worth straining much. I was just clearing out some backlog
over vacation.

Strictly up to you Jim.

Erick



On Fri, Jan 5, 2018 at 6:54 AM, David Smiley <david.w.smi...@gmail.com>
wrote:

https://issues.apache.org/jira/browse/SOLR-11809 is in progress, should be
easy and I think definitely worth backporting



On Fri, Jan 5, 2018 at 8:52 AM Adrien Grand <jpou...@gmail.com> wrote:

+1



Looking at the changelog, 7.3 has 3 bug fixes for now: LUCENE-8077,
SOLR-11783 and SOLR-11555. The Lucene change doesn't seem worth
backporting, but maybe the Solr changes should?



Le ven. 5 janv. 2018 à 12:40, jim ferenczi <jim.feren...@gmail.com> a
écrit :

Hi,

We discovered a bad bug in 7x that affects indices created in 6x with
Lucene54DocValues format. The SortedNumericDocValues created with this
format have a bug when advanceExact is used, the values retrieved for the
docs when advanceExact returns true are invalid (the pointer to the values
is not updated):

https://issues.apache.org/jira/browse/LUCENE-8117

This affects all indices created in 6x with sorted numeric doc values so I
wanted to ask if anyone objects to a bugfix release for 7.2 (7.2.1). I also
volunteer to be the release manager for this one if it is accepted.



Jim


-- 

Lucene/Solr Search Committer, Consultant, Developer, Author, Speaker

LinkedIn: http://linkedin.com/in/davidwsmiley | Book: http://www.
solrenterprisesearchserver.com

Reply via email to