On Mon, Feb 27, 2012 at 7:18 PM, Filipe David Manana <fdman...@apache.org> wrote: > Jason, can't reproduce those results, not even close: > > http://friendpaste.com/1L4pHH8WQchaLIMCWhKX9Z > > Before COUCHDB-1186 > > fdmanana 16:58:02 ~/git/hub/slow_couchdb (master)> docs=500000 > batch=50000 ./bench.sh small_doc.tpl > Server: CouchDB/1.2.0a-a68a792-git (Erlang OTP/R14B03) > {"couchdb":"Welcome","version":"1.2.0a-a68a792-git"} > > [INFO] Created DB named `db1' > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > Building view. > {"total_rows":500000,"offset":0,"rows":[ > {"id":"doc1","key":1,"value":1} > ]} > > real 0m56.241s > user 0m0.006s > sys 0m0.005s > > > After COUCHDB-1186 > > fdmanana 17:02:02 ~/git/hub/slow_couchdb (master)> docs=500000 > batch=50000 ./bench.sh small_doc.tpl > Server: CouchDB/1.2.0a-f023052-git (Erlang OTP/R14B03) > {"couchdb":"Welcome","version":"1.2.0a-f023052-git"} > > [INFO] Created DB named `db1' > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > Building view. > {"total_rows":500000,"offset":0,"rows":[ > {"id":"doc1","key":1,"value":1} > ]} > > real 1m11.694s > user 0m0.006s > sys 0m0.005s > fdmanana 17:06:01 ~/git/hub/slow_couchdb (master)> > > > 1.2.0a-f023052-git with patch > http://friendpaste.com/178nPFgfyyeGf2vtNRpL0w applied on top > > fdmanana 17:06:53 ~/git/hub/slow_couchdb (master)> docs=500000 > batch=50000 ./bench.sh small_doc.tpl > Server: CouchDB/1.2.0a-f023052-git (Erlang OTP/R14B03) > {"couchdb":"Welcome","version":"1.2.0a-f023052-git"} > > [INFO] Created DB named `db1' > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > [INFO] Uploaded 50000 documents via _bulk_docs > Building view. > {"total_rows":500000,"offset":0,"rows":[ > {"id":"doc1","key":1,"value":1} > ]} > > real 0m51.089s > user 0m0.006s > sys 0m0.004s > fdmanana 17:10:29 ~/git/hub/slow_couchdb (master)> > > > Can you try with R14B0x and also with the patch > http://friendpaste.com/178nPFgfyyeGf2vtNRpL0w ? > > Back then I made all testing on a machine with a spinning disk, so the > writer process was slower and likely dequeing more KV pairs from the > work queue on each dequeue operation. The tests I did just now are on > a machine with a ssd disk. >
Yeah, I've seen the btree behave quite differently on SSD's vs HDD's (same code had drastically different runtime characteristics). In other words, can we get a report of what type of disk everyone is running on? > > On Mon, Feb 27, 2012 at 4:40 PM, Jason Smith <j...@apache.org> wrote: >> Hi, Filipe. Most people seem to be holding their OTP build constant >> for these tests. >> >> If you have the time, would you please check out >> https://github.com/jhs/slow_couchdb >> >> It uses seatoncouch mixed with Bob's script to run a basic benchmark. >> I expect more template types to grow to help create different data >> profiles. >> >> Anyway, here are my results with 500k documents. Note that I built >> from your optimization commit, then its parent. >> >> https://gist.github.com/1928169 >> >> tl;dr = 2:50 before your commit; 4:13 after. >> >> On Mon, Feb 27, 2012 at 11:33 PM, Filipe David Manana >> <fdman...@apache.org> wrote: >>> I just tried Jason's script (modified it to use 500 000 docs instead >>> of 50 000) against 1.2.x and 1.1.1, using OTP R14B03. Here's my >>> results: >>> >>> 1.2.x: >>> >>> $ port=5984 ./test.sh >>> "none" >>> Filling db. >>> done >>> HTTP/1.1 200 OK >>> Server: CouchDB/1.2.0 (Erlang OTP/R14B03) >>> Date: Mon, 27 Feb 2012 16:08:43 GMT >>> Content-Type: text/plain; charset=utf-8 >>> Content-Length: 252 >>> Cache-Control: must-revalidate >>> >>> {"db_name":"db1","doc_count":500001,"doc_del_count":0,"update_seq":500001,"purge_seq":0,"compact_running":false,"disk_size":130494577,"data_size":130490673,"instance_start_time":"1330358830830086","disk_format_version":6,"committed_update_seq":500001} >>> Building view. >>> >>> real 1m5.725s >>> user 0m0.006s >>> sys 0m0.005s >>> done >>> >>> >>> 1.1.1: >>> >>> $ port=5984 ./test.sh >>> "" >>> Filling db. >>> done >>> HTTP/1.1 200 OK >>> Server: CouchDB/1.1.2a785d32f-git (Erlang OTP/R14B03) >>> Date: Mon, 27 Feb 2012 16:15:33 GMT >>> Content-Type: text/plain;charset=utf-8 >>> Content-Length: 230 >>> Cache-Control: must-revalidate >>> >>> {"db_name":"db1","doc_count":500001,"doc_del_count":0,"update_seq":500001,"purge_seq":0,"compact_running":false,"disk_size":122142818,"instance_start_time":"1330359233327316","disk_format_version":5,"committed_update_seq":500001} >>> Building view. >>> >>> real 1m4.249s >>> user 0m0.006s >>> sys 0m0.005s >>> done >>> >>> >>> I don't see any significant difference there. >>> >>> Regarding COUCHDB-1186, the only thing that might cause some non >>> determinism and affect performance is the queing/dequeing. Depending >>> on timings, it's possible the writer is dequeing less items per >>> dequeue operation and therefore inserting smaller batches into the >>> btree. The following small change ensures larger batches (while still >>> respecting the queue max size/item count): >>> >>> http://friendpaste.com/178nPFgfyyeGf2vtNRpL0w >>> >>> Running the test with this change: >>> >>> $ port=5984 ./test.sh >>> "none" >>> Filling db. >>> done >>> HTTP/1.1 200 OK >>> Server: CouchDB/1.2.0 (Erlang OTP/R14B03) >>> Date: Mon, 27 Feb 2012 16:23:20 GMT >>> Content-Type: text/plain; charset=utf-8 >>> Content-Length: 252 >>> Cache-Control: must-revalidate >>> >>> {"db_name":"db1","doc_count":500001,"doc_del_count":0,"update_seq":500001,"purge_seq":0,"compact_running":false,"disk_size":130494577,"data_size":130490673,"instance_start_time":"1330359706846104","disk_format_version":6,"committed_update_seq":500001} >>> Building view. >>> >>> real 0m49.762s >>> user 0m0.006s >>> sys 0m0.005s >>> done >>> >>> >>> If there's no objection, I'll push that patch. >>> >>> Also, another note, I noticed sometime ago that with master, using OTP >>> R15B I got a performance drop of 10% to 15% compared to using master >>> with OTP R14B04. Maybe it applies to 1.2.x as well. >>> >>> >>> On Mon, Feb 27, 2012 at 5:33 AM, Robert Newson <rnew...@apache.org> wrote: >>>> Bob D, can you give more details on the data set you're testing? >>>> Number of docs, size/complexity of docs, etc? Basically, enough info >>>> that I could write a script to automate building an equivalent >>>> database. >>>> >>>> I wrote a quick bash script to make a database and time a view build >>>> here: http://friendpaste.com/7kBiKJn3uX1KiGJAFPv4nK >>>> >>>> B. >>>> >>>> On 27 February 2012 13:15, Jan Lehnardt <j...@apache.org> wrote: >>>>> >>>>> On Feb 27, 2012, at 12:58 , Bob Dionne wrote: >>>>> >>>>>> Thanks for the clarification. I hope I'm not conflating things by >>>>>> continuing the discussion here, I thought that's what you requested? >>>>> >>>>> The discussion we had on IRC was regarding collecting more data items for >>>>> the performance regression before we start to draw conclusions. >>>>> >>>>> My intention here is to understand what needs doing before we can release >>>>> 1.2.0. >>>>> >>>>> I'll reply inline for the other issues. >>>>> >>>>>> I just downloaded the release candidate again to start fresh. "make >>>>>> distcheck" hangs on this step: >>>>>> >>>>>> /Users/bitdiddle/Downloads/apache-couchdb-1.2.0/apache-couchdb-1.2.0/_build/../test/etap/150-invalid-view-seq.t >>>>>> ......... 6/? >>>>>> >>>>>> Just stops completely. This is on R15B which has been rebuilt to use the >>>>>> recommended older SSL version. I haven't looked into this crashing too >>>>>> closely but I'm suspicious that I only see it with couchdb and never >>>>>> with bigcouch and never using the 1.2.x branch from source or any branch >>>>>> for that matter >>>>> >>>>> From the release you should run `make check`, not make distcheck. But I >>>>> assume you see a hang there too, as I have and others (yet not >>>>> everybody), too. I can't comment on BigCouch and what is different there. >>>>> It is interesting that 1.2.x won't hang. For me, `make check` in 1.2.x on >>>>> R15B hangs sometimes, in different places. I'm currently trying to gather >>>>> more information about this. >>>>> >>>>> The question here is whether `make check` passing in R15B is a release >>>>> requirement. In my vote I considered no, but I am happy to go with a >>>>> community decision if it emerges. What is your take here? >>>>> >>>>> In addition, this just shouldn't be a question, so we should investigate >>>>> why this happens at all and address the issue, hence COUCHDB-1424. Any >>>>> insight here would be appreciated as well. >>>>> >>>>> >>>>>> In the command line tests, 2,7, 27, and 32 fail. but it differs from run >>>>>> to run. >>>>> >>>>> I assume you mean the JS tests. Again, this isn't supposed to work in >>>>> 1.2.x. I'm happy to backport my changes from master to 1.2.x to make that >>>>> work, but I refrained from that because I didn't want to bring too much >>>>> change to a release branch. I'm happy to reconsider, but I don't think a >>>>> release vote is a good place to discuss feature backports. >>>>> >>>>> >>>>>> On Chrome attachment_ranges fails and it hangs on replicator_db >>>>> >>>>> This one is an "explaining away", but I think it is warranted. Chrome is >>>>> broken for attachment_ranges. I don't know if we reported this upstream >>>>> (Robert N?), but this isn't a release blocker. For the replicator_db >>>>> test, can you try running that in other browsers. I understand it is not >>>>> the best of situation (hence the move to the cli test suite for master), >>>>> but if you get this test to pass in at least one other browsers, this >>>>> isn't a problem that holds 1.2.x. >>>>> >>>>> >>>>>> With respect to performance I think comparisons with 1.1.x are >>>>>> important. I think almost any use case, contrived or otherwise should >>>>>> not be dismissed as a pathological or edge case. Bob's script is as >>>>>> simple as it gets and to me is a great smoke test. We need to figure out >>>>>> the reason 1.2 is clearly slower in this case. If there are specific >>>>>> scenarios that 1.2.x is optimized for then we should document that and >>>>>> provide reasons for the trade-offs >>>>> >>>>> I want to make absolutely clear that I take any report of performance >>>>> regression very seriously. But I'm rather annoyed that no information >>>>> about this ends up on dev@. I understand that on IRC there's some shared >>>>> understanding of a few scenarios where performance regressions can be >>>>> shown. I asked three times now that these be posted to this mailing list. >>>>> I'm not asking for a comprehensive report, but anything really. I found >>>>> Robert Newson's simple test script on IRC and ran that to test a >>>>> suspicion of mine which I posted in an earlier mail (tiny docs -> slower, >>>>> bigger docs -> faster). Nobody else bothered to post this here. I see no >>>>> discussion about what is observed, what is expected, what would be >>>>> acceptable for a release of 1.2.0 as is and what not. >>>>> >>>>> As far as this list is concerned, we know that a few people claimed that >>>>> things are slower and it's very real and that we should hold the 1.2.0 >>>>> release for it. I'm more than happy to hold the release until we figured >>>>> out the things I asked for above and help out figuring it all out. But we >>>>> need something to work with here. >>>>> >>>>> I also understand that this is a voluntary project and people don't have >>>>> infinite time to spend, but at least a message of "we're collecting >>>>> things, will report when done", would be *great* to start. So far we only >>>>> have a "hold the horses, there might be a something going on". >>>>> >>>>> Please let me know if this request is unreasonable or whether I am >>>>> overreacting. >>>>> >>>>> Sorry for the rant. >>>>> >>>>> To anyone who has been looking into performance regression, can you >>>>> please send to this list any info you have? If you have a comprehensive >>>>> analysis, awesome, if you just ran some script on a machine, just send us >>>>> that, let's collect all the data to get this situation solved! We need >>>>> your help. >>>>> >>>>> >>>>> tl;dr: >>>>> >>>>> There's three issues at hand: >>>>> >>>>> - Robert D -1'd a release artefact. We want to understand what needs to >>>>> happen to make a release. This includes assessing the issues he raises >>>>> and squaring them against the release vote. >>>>> >>>>> - There's a vague (as far as dev@ is concerned) report about a >>>>> performance regression. We need to get behind that. >>>>> >>>>> - There's been a non-dev@ discussion about the performance regression >>>>> and that is referenced to influence a dev@ decision. We need that >>>>> discussion's information on dev@ to proceed. >>>>> >>>>> >>>>> And to make it absolutely clear again. The performance regression *is* an >>>>> issue and I am very grateful for the people, including Robert Newson, >>>>> Robert Dionne and Jason Smith, who look into it. It's just that we need >>>>> to treat this as an issue and get all this info onto dev@ or into JRIA. >>>>> >>>>> >>>>> Cheers >>>>> Jan >>>>> -- >>>>> >>>>> >>>>> >>>>>> >>>>>> Cheers, >>>>>> >>>>>> Bob >>>>>> >>>>>> >>>>>> On Feb 26, 2012, at 4:07 PM, Jan Lehnardt wrote: >>>>>> >>>>>>> Bob, >>>>>>> >>>>>>> thanks for your reply >>>>>>> >>>>>>> I wasn't implying we should try to explain anything away. All of these >>>>>>> are valid concerns, I just wanted to get a better understanding on >>>>>>> where the bit flips from +0 to -1 and subsequently, how to address that >>>>>>> boundary. >>>>>> >>>>>> >>>>>> >>>>>> >>>>>>> Ideally we can just fix all of the things you mention, but I think it >>>>>>> is important to understand them in detail, that's why I was going into >>>>>>> them. Ultimately, I want to understand what we need to do to ship 1.2.0. >>>>>>> >>>>>>> On Feb 26, 2012, at 21:22 , Bob Dionne wrote: >>>>>>> >>>>>>>> Jan, >>>>>>>> >>>>>>>> I'm -1 based on all of my evaluation. I've spent a few hours on this >>>>>>>> release now yesterday and today. It doesn't really pass what I would >>>>>>>> call the "smoke test". Almost everything I've run into has an >>>>>>>> explanation: >>>>>>>> >>>>>>>> 1. crashes out of the box - that's R15B, you need to recompile SSL and >>>>>>>> Erlang (we'll note on release notes) >>>>>>> >>>>>>> Have we spent any time on figuring out what the trouble here is? >>>>>>> >>>>>>> >>>>>>>> 2. etaps hang running make check. Known issue. Our etap code is out of >>>>>>>> date, recent versions of etap don't even run their own unit tests >>>>>>> >>>>>>> I have seen the etap hang as well, and I wasn't diligent enough to >>>>>>> report it in JIRA, I have done so now (COUCHDB-1424). >>>>>>> >>>>>>> >>>>>>>> 3. Futon tests fail. Some are known bugs (attachment ranges in Chrome) >>>>>>>> . Both Chrome and Safari also hang >>>>>>> >>>>>>> Do you have more details on where Chrome and Safari hang? Can you try >>>>>>> their private browsing features, double/triple check that caches are >>>>>>> empty? Can you get to a situation where you get all tests succeeding >>>>>>> across all browsers, even if individual ones fail on one or two others? >>>>>>> >>>>>>> >>>>>>>> 4. standalone JS tests fail. Again most of these run when run by >>>>>>>> themselves >>>>>>> >>>>>>> Which ones? >>>>>>> >>>>>>> >>>>>>>> 5. performance. I used real production data *because* Stefan on user >>>>>>>> reported performance degradation on his data set. Any numbers are >>>>>>>> meaningless for a single test. I also ran scripts that BobN and Jason >>>>>>>> Smith posted that show a difference between 1.1.x and 1.2.x >>>>>>> >>>>>>> You are conflating an IRC discussion we've had into this thread. The >>>>>>> performance regression reported is a good reason to look into other >>>>>>> scenarios where we can show slowdowns. But we need to understand what's >>>>>>> happening. Just from looking at dev@ all I see is some handwaving about >>>>>>> some reports some people have done (Not to discourage any work that has >>>>>>> been done on IRC and user@, but for the sake of a release vote thread, >>>>>>> this related information needs to be on this mailing list). >>>>>>> >>>>>>> As I said on IRC, I'm happy to get my hands dirty to understand the >>>>>>> regression at hand. But we need to know where we'd draw a line and say >>>>>>> this isn't acceptable for a 1.2.0. >>>>>>> >>>>>>> >>>>>>>> 6. Reviewed patch pointed to by Jason that may be the cause but it's >>>>>>>> hard to say without knowing the code analysis that went into the >>>>>>>> changes. You can see obvious local optimizations that make good sense >>>>>>>> but those are often the ones that get you, without knowing the call >>>>>>>> counts. >>>>>>> >>>>>>> That is a point that wasn't included in your previous mail. It's great >>>>>>> that there is progress, thanks for looking into this! >>>>>>> >>>>>>> >>>>>>>> Many of these issues can be explained away, but I think end users will >>>>>>>> be less forgiving. I think we already struggle with view performance. >>>>>>>> I'm interested to see how others evaluate this regression. >>>>>>>> I'll try this seatoncouch tool you mention later to see if I can >>>>>>>> construct some more definitive tests. >>>>>>> >>>>>>> Again, I'm not trying to explain anything away. I want to get a shared >>>>>>> understanding of the issues you raised and where we stand on solving >>>>>>> them squared against the ongoing 1.2.0 release. >>>>>>> >>>>>>> And again: Thanks for doing this thorough review and looking into >>>>>>> performance issue. I hope with your help we can understand all these >>>>>>> things a lot better very soon :) >>>>>>> >>>>>>> Cheers >>>>>>> Jan >>>>>>> -- >>>>>>> >>>>>>> >>>>>>>> >>>>>>>> Best, >>>>>>>> >>>>>>>> Bob >>>>>>>> On Feb 26, 2012, at 2:29 PM, Jan Lehnardt wrote: >>>>>>>> >>>>>>>>> >>>>>>>>> On Feb 26, 2012, at 13:58 , Bob Dionne wrote: >>>>>>>>> >>>>>>>>>> -1 >>>>>>>>>> >>>>>>>>>> R15B on OS X Lion >>>>>>>>>> >>>>>>>>>> I rebuilt OTP with an older SSL and that gets past all the crashes >>>>>>>>>> (thanks Filipe). I still see hangs when running make check, though >>>>>>>>>> any particular etap that hangs will run ok by itself. The Futon >>>>>>>>>> tests never run to completion in Chrome without hanging and the >>>>>>>>>> standalone JS tests also have fails. >>>>>>>>> >>>>>>>>> What part of this do you consider the -1? Can you try running the JS >>>>>>>>> tests in Firefox and or Safari? Can you get all tests pass at least >>>>>>>>> once across all browsers? The cli JS suite isn't supposed to work, so >>>>>>>>> that isn't a criterion. I've seen the hang in make check for R15B >>>>>>>>> while individual tests run as well, but I don't consider this >>>>>>>>> blocking. While I understand and support the notion that tests >>>>>>>>> shouldn't fail, period, we gotta work with what we have and master >>>>>>>>> already has significant improvements. What would you like to see >>>>>>>>> changed to not -1 this release? >>>>>>>>> >>>>>>>>>> I tested the performance of view indexing, using a modest 200K doc >>>>>>>>>> db with a large complex view and there's a clear regression between >>>>>>>>>> 1.1.x and 1.2.x Others report similar results >>>>>>>>> >>>>>>>>> What is a large complex view? The complexity of the map/reduce >>>>>>>>> functions is rarely an indicator of performance, it's usually input >>>>>>>>> doc size and output/emit()/reduce data size. How big are the docs in >>>>>>>>> your test and how big is the returned data? I understand the changes >>>>>>>>> for 1.2.x will improve larger-data scenarios more significantly. >>>>>>>>> >>>>>>>>> Cheers >>>>>>>>> Jan >>>>>>>>> -- >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Feb 23, 2012, at 5:25 PM, Bob Dionne wrote: >>>>>>>>>> >>>>>>>>>>> sorry Noah, I'm in debug mode today so I don't care to start >>>>>>>>>>> mucking with my stack, recompiling erlang, etc... >>>>>>>>>>> >>>>>>>>>>> I did try using that build repeatedly and it crashes all the time. >>>>>>>>>>> I find it very odd and I had seen those before as I said on my >>>>>>>>>>> older macbook. >>>>>>>>>>> >>>>>>>>>>> I do see the hangs Jan describes in the etaps, they have been there >>>>>>>>>>> right along, so I'm confident this just the SSL issue. Why it only >>>>>>>>>>> happens on the build is puzzling, any source build of any branch >>>>>>>>>>> works just peachy. >>>>>>>>>>> >>>>>>>>>>> So I'd say I'm +1 based on my use of the 1.2.x branch but I'd like >>>>>>>>>>> to hear from Stefan, who reported the severe performance >>>>>>>>>>> regression. BobN seems to think we can ignore that, it's something >>>>>>>>>>> flaky in that fellow's environment. I tend to agree but I'm >>>>>>>>>>> conservative >>>>>>>>>>> >>>>>>>>>>> On Feb 23, 2012, at 1:23 PM, Noah Slater wrote: >>>>>>>>>>> >>>>>>>>>>>> Can someone convince me this bus error stuff and segfaults is not a >>>>>>>>>>>> blocking issue. >>>>>>>>>>>> >>>>>>>>>>>> Bob tells me that he's followed the steps above and he's still >>>>>>>>>>>> experiencing >>>>>>>>>>>> the issues. >>>>>>>>>>>> >>>>>>>>>>>> Bob, you did follow the steps to install your own SSL right? >>>>>>>>>>>> >>>>>>>>>>>> On Thu, Feb 23, 2012 at 5:09 PM, Jan Lehnardt <j...@apache.org> >>>>>>>>>>>> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Feb 23, 2012, at 00:28 , Noah Slater wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Hello, >>>>>>>>>>>>>> >>>>>>>>>>>>>> I would like call a vote for the Apache CouchDB 1.2.0 release, >>>>>>>>>>>>>> second >>>>>>>>>>>>> round. >>>>>>>>>>>>>> >>>>>>>>>>>>>> We encourage the whole community to download and test these >>>>>>>>>>>>>> release artifacts so that any critical issues can be resolved >>>>>>>>>>>>>> before the >>>>>>>>>>>>>> release is made. Everyone is free to vote on this release, so >>>>>>>>>>>>>> get stuck >>>>>>>>>>>>> in! >>>>>>>>>>>>>> >>>>>>>>>>>>>> We are voting on the following release artifacts: >>>>>>>>>>>>>> >>>>>>>>>>>>>> http://people.apache.org/~nslater/dist/1.2.0/ >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> These artifacts have been built from the following tree-ish in >>>>>>>>>>>>>> Git: >>>>>>>>>>>>>> >>>>>>>>>>>>>> 4cd60f3d1683a3445c3248f48ae064fb573db2a1 >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> Please follow the test procedure before voting: >>>>>>>>>>>>>> >>>>>>>>>>>>>> http://wiki.apache.org/couchdb/Test_procedure >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> Thank you. >>>>>>>>>>>>>> >>>>>>>>>>>>>> Happy voting, >>>>>>>>>>>>> >>>>>>>>>>>>> Signature and hashes check out. >>>>>>>>>>>>> >>>>>>>>>>>>> Mac OS X 10.7.3, 64bit, SpiderMonkey 1.8.0, Erlang R14B04: make >>>>>>>>>>>>> check >>>>>>>>>>>>> works fine, browser tests in Safari work fine. >>>>>>>>>>>>> >>>>>>>>>>>>> Mac OS X 10.7.3, 64bit, SpiderMonkey 1.8.5, Erlang R14B04: make >>>>>>>>>>>>> check >>>>>>>>>>>>> works fine, browser tests in Safari work fine. >>>>>>>>>>>>> >>>>>>>>>>>>> FreeBSD 9.0, 64bit, SpiderMonkey 1.7.0, Erlang R14B04: make check >>>>>>>>>>>>> works >>>>>>>>>>>>> fine, browser tests in Safari work fine. >>>>>>>>>>>>> >>>>>>>>>>>>> CentOS 6.2, 64bit, SpiderMonkey 1.8.5, Erlang R14B04: make check >>>>>>>>>>>>> works >>>>>>>>>>>>> fine, browser tests in Firefox work fine. >>>>>>>>>>>>> >>>>>>>>>>>>> Ubuntu 11.4, 64bit, SpiderMonkey 1.8.5, Erlang R14B02: make check >>>>>>>>>>>>> works >>>>>>>>>>>>> fine, browser tests in Firefox work fine. >>>>>>>>>>>>> >>>>>>>>>>>>> Ubuntu 10.4, 32bit, SpiderMonkey 1.8.0, Erlang R13B03: make check >>>>>>>>>>>>> fails in >>>>>>>>>>>>> - 076-file-compression.t: https://gist.github.com/1893373 >>>>>>>>>>>>> - 220-compaction-daemon.t: https://gist.github.com/1893387 >>>>>>>>>>>>> This on runs in a VM and is 32bit, so I don't know if there's >>>>>>>>>>>>> anything in >>>>>>>>>>>>> the tests that rely on 64bittyness or the R14B03. Filipe, I think >>>>>>>>>>>>> you >>>>>>>>>>>>> worked on both features, do you have an idea? >>>>>>>>>>>>> >>>>>>>>>>>>> I tried running it all through Erlang R15B on Mac OS X 1.7.3, but >>>>>>>>>>>>> a good >>>>>>>>>>>>> way into `make check` the tests would just stop and hang. The >>>>>>>>>>>>> last time, >>>>>>>>>>>>> repeatedly in 160-vhosts.t, but when run alone, that test >>>>>>>>>>>>> finished in under >>>>>>>>>>>>> five seconds. I'm not sure what the issue is here. >>>>>>>>>>>>> >>>>>>>>>>>>> Despite the things above, I'm happy to give this a +1 if we put a >>>>>>>>>>>>> warning >>>>>>>>>>>>> about R15B on the download page. >>>>>>>>>>>>> >>>>>>>>>>>>> Great work all! >>>>>>>>>>>>> >>>>>>>>>>>>> Cheers >>>>>>>>>>>>> Jan >>>>>>>>>>>>> -- >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>> >>> >>> >>> -- >>> Filipe David Manana, >>> >>> "Reasonable men adapt themselves to the world. >>> Unreasonable men adapt the world to themselves. >>> That's why all progress depends on unreasonable men." > > > > -- > Filipe David Manana, > > "Reasonable men adapt themselves to the world. > Unreasonable men adapt the world to themselves. > That's why all progress depends on unreasonable men."