"you might want to revisit the release process"

You mean "we" here. :)

B.


On 29 February 2012 16:23, Jan Lehnardt <j...@apache.org> wrote:
> Hi Bob,
>
> thanks for your reply. I think this is a bit of a convoluted situation now 
> and I want to apologise for anything that might have upset you or anyone 
> here. I didn't mean to be insulting or anything.
>
> I think the 1.2.0 release is very important for this community and the 
> project, so I am very interested in getting it out. I'm not interested in 
> getting it out at all costs but I do want to clarify what's needed to do the 
> release.
>
> In the process, your -1 turned out to trigger a hold on the 1.2.0 release, 
> even though you write you didn't intend that. Also in the process I confused 
> your -1 to ask for the hold and as a result I got annoyed that you didn't 
> provide more information. I also asked for clarification on the other issues 
> you raised (e.g. running the test suite in Firefox/across multiple browsers) 
> and you haven't responded to yet (I understand time constraints, but I'd have 
> appreciated a "gonna work on it" sort of quick response).
>
> I'm sorry I got annoyed and let it out on you.
>
> All I want is to understand what it takes to release 1.2.0 and we need all 
> your (and everybody's) support to get there. So thanks for any contributions 
> to this discussion.
>
> Cheers
> Jan
> --
>
>
> On Feb 29, 2012, at 16:27 , Bob Dionne wrote:
>
>> Jan,
>>
>> Sorry it took me a while to respond to this. As I said the db I was testing 
>> with is about ~200K docs, with one large ddoc, about 10 views. It takes a 
>> considerable long time to index, both on 1.1.x and 1.2.x but on 1.2.x it had 
>> not made the same amount of progress after 45 minutes on 1.2.x as did the 
>> 1.1.x run after 20 minutes, so I assumed that something has slowed, 40% 
>> maybe, that's likely a number I pulled out of my hat from the chat and 
>> reports of others.
>>
>> I understand this is hand-wavy, which is why I was reluctant to jump to 
>> conclusions. Please note that I did not state we should stop this release, 
>> nor did I state that any of the issues I enumerated were showstoppers. I 
>> stated, with a -1 vote that the overall experience led me to conclude the 
>> release was not sound. It's one vote and I always am happy to go with the 
>> majority.
>>
>> I thought we were off to a good start on Sunday when I was digging in here. 
>> Yes it's off topic as you and Noah agreed, but once you've pushed this out 
>> you might want to revisit the release process. This slow_couchdb is 
>> definitely a good smoke test, I like it, but beyond that it's sort of an 
>> instance of "You're doing it all wrong" when it comes to performance. I'll 
>> be happy to elaborate on this when we get to that.
>>
>> Best Regards,
>>
>> Bob
>>
>>
>> On Feb 27, 2012, at 8:15 AM, Jan Lehnardt wrote:
>>
>>>
>>> On Feb 27, 2012, at 12:58 , Bob Dionne wrote:
>>>
>>>> Thanks for the clarification. I hope I'm not conflating things by 
>>>> continuing the discussion here, I thought that's what you requested?
>>>
>>> The discussion we had on IRC was regarding collecting more data items for 
>>> the performance regression before we start to draw conclusions.
>>>
>>> My intention here is to understand what needs doing before we can release 
>>> 1.2.0.
>>>
>>> I'll reply inline for the other issues.
>>>
>>>> I just downloaded the release candidate again to start fresh. "make 
>>>> distcheck" hangs on this step:
>>>>
>>>> /Users/bitdiddle/Downloads/apache-couchdb-1.2.0/apache-couchdb-1.2.0/_build/../test/etap/150-invalid-view-seq.t
>>>>  ......... 6/?
>>>>
>>>> Just stops completely. This is on R15B which has been rebuilt to use the 
>>>> recommended older SSL version. I haven't looked into this crashing too 
>>>> closely but I'm suspicious that I only see it with couchdb and never with 
>>>> bigcouch and never using the 1.2.x branch from source or any branch for 
>>>> that matter
>>>
>>> From the release you should run `make check`, not make distcheck. But I 
>>> assume you see a hang there too, as I have and others (yet not everybody), 
>>> too. I can't comment on BigCouch and what is different there. It is 
>>> interesting that 1.2.x won't hang. For me, `make check` in 1.2.x on R15B 
>>> hangs sometimes, in different places. I'm currently trying to gather more 
>>> information about this.
>>>
>>> The question here is whether `make check` passing in R15B is a release 
>>> requirement. In my vote I considered no, but I am happy to go with a 
>>> community decision if it emerges. What is your take here?
>>>
>>> In addition, this just shouldn't be a question, so we should investigate 
>>> why this happens at all and address the issue, hence COUCHDB-1424. Any 
>>> insight here would be appreciated as well.
>>>
>>>
>>>> In the command line tests, 2,7, 27, and 32 fail. but it differs from run 
>>>> to run.
>>>
>>> I assume you mean the JS tests. Again, this isn't supposed to work in 
>>> 1.2.x. I'm happy to backport my changes from master to 1.2.x to make that 
>>> work, but I refrained from that because I didn't want to bring too much 
>>> change to a release branch. I'm happy to reconsider, but I don't think a 
>>> release vote is a good place to discuss feature backports.
>>>
>>>
>>>> On Chrome attachment_ranges fails and it hangs on replicator_db
>>>
>>> This one is an "explaining away", but I think it is warranted. Chrome is 
>>> broken for attachment_ranges. I don't know if we reported this upstream 
>>> (Robert N?), but this isn't a release blocker. For the replicator_db test, 
>>> can you try running that in other browsers. I understand it is not the best 
>>> of situation (hence the move to the cli test suite for master), but if you 
>>> get this test to pass in at least one other browsers, this isn't a problem 
>>> that holds 1.2.x.
>>>
>>>
>>>> With respect to performance I think comparisons with 1.1.x are important. 
>>>> I think almost any use case, contrived or otherwise should not be 
>>>> dismissed as a pathological or edge case. Bob's script is as simple as it 
>>>> gets and to me is a great smoke test. We need to figure out the reason 1.2 
>>>> is clearly slower in this case. If there are specific scenarios that 1.2.x 
>>>> is optimized for then we should document that and provide reasons for the 
>>>> trade-offs
>>>
>>> I want to make absolutely clear that I take any report of performance 
>>> regression very seriously. But I'm rather annoyed that no information about 
>>> this ends up on dev@. I understand that on IRC there's some shared 
>>> understanding of a few scenarios where performance regressions can be 
>>> shown. I asked three times now that these be posted to this mailing list. 
>>> I'm not asking for a comprehensive report, but anything really. I found 
>>> Robert Newson's simple test script on IRC and ran that to test a suspicion 
>>> of mine which I posted in an earlier mail (tiny docs -> slower, bigger docs 
>>> -> faster). Nobody else bothered to post this here. I see no discussion 
>>> about what is observed, what is expected, what would be acceptable for a 
>>> release of 1.2.0 as is and what not.
>>>
>>> As far as this list is concerned, we know that a few people claimed that 
>>> things are slower and it's very real and that we should hold the 1.2.0 
>>> release for it. I'm more than happy to hold the release until we figured 
>>> out the things I asked for above and help out figuring it all out. But we 
>>> need something to work with here.
>>>
>>> I also understand that this is a voluntary project and people don't have 
>>> infinite time to spend, but at least a message of "we're collecting things, 
>>> will report when done", would be *great* to start. So far we only have a 
>>> "hold the horses, there might be a something going on".
>>>
>>> Please let me know if this request is unreasonable or whether I am 
>>> overreacting.
>>>
>>> Sorry for the rant.
>>>
>>> To anyone who has been looking into performance regression, can you please 
>>> send to this list any info you have? If you have a comprehensive analysis, 
>>> awesome, if you just ran some script on a machine, just send us that, let's 
>>> collect all the data to get this situation solved! We need your help.
>>>
>>>
>>> tl;dr:
>>>
>>> There's three issues at hand:
>>>
>>> - Robert D -1'd a release artefact. We want to understand what needs to 
>>> happen to make a release. This includes assessing the issues he raises and 
>>> squaring them against the release vote.
>>>
>>> - There's a vague (as far as dev@ is concerned) report about a performance 
>>> regression. We need to get behind that.
>>>
>>> - There's been a non-dev@ discussion about the performance regression and 
>>> that is referenced to influence a dev@ decision. We need that discussion's 
>>> information on dev@ to proceed.
>>>
>>>
>>> And to make it absolutely clear again. The performance regression *is* an 
>>> issue and I am very grateful for the people, including Robert Newson, 
>>> Robert Dionne and Jason Smith, who look into it. It's just that we need to 
>>> treat this as an issue and get all this info onto dev@ or into JRIA.
>>>
>>>
>>> Cheers
>>> Jan
>>> --
>>>
>>>
>>>
>>>>
>>>> Cheers,
>>>>
>>>> Bob
>>>>
>>>>
>>>> On Feb 26, 2012, at 4:07 PM, Jan Lehnardt wrote:
>>>>
>>>>> Bob,
>>>>>
>>>>> thanks for your reply
>>>>>
>>>>> I wasn't implying we should try to explain anything away. All of these 
>>>>> are valid concerns, I just wanted to get a better understanding on where 
>>>>> the bit flips from +0 to -1 and subsequently, how to address that 
>>>>> boundary.
>>>>
>>>>
>>>>
>>>>
>>>>> Ideally we can just fix all of the things you mention, but I think it is 
>>>>> important to understand them in detail, that's why I was going into them. 
>>>>> Ultimately, I want to understand what we need to do to ship 1.2.0.
>>>>>
>>>>> On Feb 26, 2012, at 21:22 , Bob Dionne wrote:
>>>>>
>>>>>> Jan,
>>>>>>
>>>>>> I'm -1 based on all of my evaluation. I've spent a few hours on this 
>>>>>> release now yesterday and today. It doesn't really pass what I would 
>>>>>> call the "smoke test". Almost everything I've run into has an 
>>>>>> explanation:
>>>>>>
>>>>>> 1. crashes out of the box - that's R15B, you need to recompile SSL and 
>>>>>> Erlang (we'll note on release notes)
>>>>>
>>>>> Have we spent any time on figuring out what the trouble here is?
>>>>>
>>>>>
>>>>>> 2. etaps hang running make check. Known issue. Our etap code is out of 
>>>>>> date, recent versions of etap don't even run their own unit tests
>>>>>
>>>>> I have seen the etap hang as well, and I wasn't diligent enough to report 
>>>>> it in JIRA, I have done so now (COUCHDB-1424).
>>>>>
>>>>>
>>>>>> 3. Futon tests fail. Some are known bugs (attachment ranges in Chrome) . 
>>>>>> Both Chrome and Safari also hang
>>>>>
>>>>> Do you have more details on where Chrome and Safari hang? Can you try 
>>>>> their private browsing features, double/triple check that caches are 
>>>>> empty? Can you get to a situation where you get all tests succeeding 
>>>>> across all browsers, even if individual ones fail on one or two others?
>>>>>
>>>>>
>>>>>> 4. standalone JS tests fail. Again most of these run when run by 
>>>>>> themselves
>>>>>
>>>>> Which ones?
>>>>>
>>>>>
>>>>>> 5. performance. I used real production data *because* Stefan on user 
>>>>>> reported performance degradation on his data set. Any numbers are 
>>>>>> meaningless for a single test. I also ran scripts that BobN and Jason 
>>>>>> Smith posted that show a difference between 1.1.x and 1.2.x
>>>>>
>>>>> You are conflating an IRC discussion we've had into this thread. The 
>>>>> performance regression reported is a good reason to look into other 
>>>>> scenarios where we can show slowdowns. But we need to understand what's 
>>>>> happening. Just from looking at dev@ all I see is some handwaving about 
>>>>> some reports some people have done (Not to discourage any work that has 
>>>>> been done on IRC and user@, but for the sake of a release vote thread, 
>>>>> this related information needs to be on this mailing list).
>>>>>
>>>>> As I said on IRC, I'm happy to get my hands dirty to understand the 
>>>>> regression at hand. But we need to know where we'd draw a line and say 
>>>>> this isn't acceptable for a 1.2.0.
>>>>>
>>>>>
>>>>>> 6. Reviewed patch pointed to by Jason that may be the cause but it's 
>>>>>> hard to say without knowing the code analysis that went into the 
>>>>>> changes. You can see obvious local optimizations that make good sense 
>>>>>> but those are often the ones that get you, without knowing the call 
>>>>>> counts.
>>>>>
>>>>> That is a point that wasn't included in your previous mail. It's great 
>>>>> that there is progress, thanks for looking into this!
>>>>>
>>>>>
>>>>>> Many of these issues can be explained away, but I think end users will 
>>>>>> be less forgiving. I think we already struggle with view performance. 
>>>>>> I'm interested to see how others evaluate this regression.
>>>>>> I'll try this seatoncouch tool you mention later to see if I can 
>>>>>> construct some more definitive tests.
>>>>>
>>>>> Again, I'm not trying to explain anything away. I want to get a shared 
>>>>> understanding of the issues you raised and where we stand on solving them 
>>>>> squared against the ongoing 1.2.0 release.
>>>>>
>>>>> And again: Thanks for doing this thorough review and looking into 
>>>>> performance issue. I hope with your help we can understand all these 
>>>>> things a lot better very soon :)
>>>>>
>>>>> Cheers
>>>>> Jan
>>>>> --
>>>>>
>>>>>
>>>>>>
>>>>>> Best,
>>>>>>
>>>>>> Bob
>>>>>> On Feb 26, 2012, at 2:29 PM, Jan Lehnardt wrote:
>>>>>>
>>>>>>>
>>>>>>> On Feb 26, 2012, at 13:58 , Bob Dionne wrote:
>>>>>>>
>>>>>>>> -1
>>>>>>>>
>>>>>>>> R15B on OS X Lion
>>>>>>>>
>>>>>>>> I rebuilt OTP with an older SSL and that gets past all the crashes 
>>>>>>>> (thanks Filipe). I still see hangs when running make check, though any 
>>>>>>>> particular etap that hangs will run ok by itself. The Futon tests 
>>>>>>>> never run to completion in Chrome without hanging and the standalone 
>>>>>>>> JS tests also have fails.
>>>>>>>
>>>>>>> What part of this do you consider the -1? Can you try running the JS 
>>>>>>> tests in Firefox and or Safari? Can you get all tests pass at least 
>>>>>>> once across all browsers? The cli JS suite isn't supposed to work, so 
>>>>>>> that isn't a criterion. I've seen the hang in make check for R15B while 
>>>>>>> individual tests run as well, but I don't consider this blocking. While 
>>>>>>> I understand and support the notion that tests shouldn't fail, period, 
>>>>>>> we gotta work with what we have and master already has significant 
>>>>>>> improvements. What would you like to see changed to not -1 this release?
>>>>>>>
>>>>>>>> I tested the performance of view indexing, using a modest 200K doc db 
>>>>>>>> with a large complex view and there's a clear regression between 1.1.x 
>>>>>>>> and 1.2.x Others report similar results
>>>>>>>
>>>>>>> What is a large complex view? The complexity of the map/reduce 
>>>>>>> functions is rarely an indicator of performance, it's usually input doc 
>>>>>>> size and output/emit()/reduce data size. How big are the docs in your 
>>>>>>> test and how big is the returned data? I understand the changes for 
>>>>>>> 1.2.x will improve larger-data scenarios more significantly.
>>>>>>>
>>>>>>> Cheers
>>>>>>> Jan
>>>>>>> --
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>>
>>>>>>>> On Feb 23, 2012, at 5:25 PM, Bob Dionne wrote:
>>>>>>>>
>>>>>>>>> sorry Noah, I'm in debug mode today so I don't care to start mucking 
>>>>>>>>> with my stack, recompiling erlang, etc...
>>>>>>>>>
>>>>>>>>> I did try using that build repeatedly and it crashes all the time. I 
>>>>>>>>> find it very odd and I had seen those before as I said on my older 
>>>>>>>>> macbook.
>>>>>>>>>
>>>>>>>>> I do see the hangs Jan describes in the etaps, they have been there 
>>>>>>>>> right along, so I'm confident this just the SSL issue. Why it only 
>>>>>>>>> happens on the build is puzzling, any source build of any branch 
>>>>>>>>> works just peachy.
>>>>>>>>>
>>>>>>>>> So I'd say I'm +1 based on my use of the 1.2.x branch but I'd like to 
>>>>>>>>> hear from Stefan, who reported the severe performance regression. 
>>>>>>>>> BobN seems to think we can ignore that, it's something flaky in that 
>>>>>>>>> fellow's environment. I tend to agree but I'm conservative
>>>>>>>>>
>>>>>>>>> On Feb 23, 2012, at 1:23 PM, Noah Slater wrote:
>>>>>>>>>
>>>>>>>>>> Can someone convince me this bus error stuff and segfaults is not a
>>>>>>>>>> blocking issue.
>>>>>>>>>>
>>>>>>>>>> Bob tells me that he's followed the steps above and he's still 
>>>>>>>>>> experiencing
>>>>>>>>>> the issues.
>>>>>>>>>>
>>>>>>>>>> Bob, you did follow the steps to install your own SSL right?
>>>>>>>>>>
>>>>>>>>>> On Thu, Feb 23, 2012 at 5:09 PM, Jan Lehnardt <j...@apache.org> 
>>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Feb 23, 2012, at 00:28 , Noah Slater wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Hello,
>>>>>>>>>>>>
>>>>>>>>>>>> I would like call a vote for the Apache CouchDB 1.2.0 release, 
>>>>>>>>>>>> second
>>>>>>>>>>> round.
>>>>>>>>>>>>
>>>>>>>>>>>> We encourage the whole community to download and test these
>>>>>>>>>>>> release artifacts so that any critical issues can be resolved 
>>>>>>>>>>>> before the
>>>>>>>>>>>> release is made. Everyone is free to vote on this release, so get 
>>>>>>>>>>>> stuck
>>>>>>>>>>> in!
>>>>>>>>>>>>
>>>>>>>>>>>> We are voting on the following release artifacts:
>>>>>>>>>>>>
>>>>>>>>>>>> http://people.apache.org/~nslater/dist/1.2.0/
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> These artifacts have been built from the following tree-ish in Git:
>>>>>>>>>>>>
>>>>>>>>>>>> 4cd60f3d1683a3445c3248f48ae064fb573db2a1
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Please follow the test procedure before voting:
>>>>>>>>>>>>
>>>>>>>>>>>> http://wiki.apache.org/couchdb/Test_procedure
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Thank you.
>>>>>>>>>>>>
>>>>>>>>>>>> Happy voting,
>>>>>>>>>>>
>>>>>>>>>>> Signature and hashes check out.
>>>>>>>>>>>
>>>>>>>>>>> Mac OS X 10.7.3, 64bit, SpiderMonkey 1.8.0, Erlang R14B04: make 
>>>>>>>>>>> check
>>>>>>>>>>> works fine, browser tests in Safari work fine.
>>>>>>>>>>>
>>>>>>>>>>> Mac OS X 10.7.3, 64bit, SpiderMonkey 1.8.5, Erlang R14B04: make 
>>>>>>>>>>> check
>>>>>>>>>>> works fine, browser tests in Safari work fine.
>>>>>>>>>>>
>>>>>>>>>>> FreeBSD 9.0, 64bit, SpiderMonkey 1.7.0, Erlang R14B04: make check 
>>>>>>>>>>> works
>>>>>>>>>>> fine, browser tests in Safari work fine.
>>>>>>>>>>>
>>>>>>>>>>> CentOS 6.2, 64bit, SpiderMonkey 1.8.5, Erlang R14B04: make check 
>>>>>>>>>>> works
>>>>>>>>>>> fine, browser tests in Firefox work fine.
>>>>>>>>>>>
>>>>>>>>>>> Ubuntu 11.4, 64bit, SpiderMonkey 1.8.5, Erlang R14B02: make check 
>>>>>>>>>>> works
>>>>>>>>>>> fine, browser tests in Firefox work fine.
>>>>>>>>>>>
>>>>>>>>>>> Ubuntu 10.4, 32bit, SpiderMonkey 1.8.0, Erlang R13B03: make check 
>>>>>>>>>>> fails in
>>>>>>>>>>> - 076-file-compression.t: https://gist.github.com/1893373
>>>>>>>>>>> - 220-compaction-daemon.t: https://gist.github.com/1893387
>>>>>>>>>>> This on runs in a VM and is 32bit, so I don't know if there's 
>>>>>>>>>>> anything in
>>>>>>>>>>> the tests that rely on 64bittyness or the R14B03. Filipe, I think 
>>>>>>>>>>> you
>>>>>>>>>>> worked on both features, do you have an idea?
>>>>>>>>>>>
>>>>>>>>>>> I tried running it all through Erlang R15B on Mac OS X 1.7.3, but a 
>>>>>>>>>>> good
>>>>>>>>>>> way into `make check` the tests would just stop and hang. The last 
>>>>>>>>>>> time,
>>>>>>>>>>> repeatedly in 160-vhosts.t, but when run alone, that test finished 
>>>>>>>>>>> in under
>>>>>>>>>>> five seconds. I'm not sure what the issue is here.
>>>>>>>>>>>
>>>>>>>>>>> Despite the things above, I'm happy to give this a +1 if we put a 
>>>>>>>>>>> warning
>>>>>>>>>>> about R15B on the download page.
>>>>>>>>>>>
>>>>>>>>>>> Great work all!
>>>>>>>>>>>
>>>>>>>>>>> Cheers
>>>>>>>>>>> Jan
>>>>>>>>>>> --
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to