After more investigation, the ConnectionRefused exception could be reproduced with "mvn -Dtest=<case_name> test" after a complete run of all cases through "mvn -PrunAllTests clean test", but cannot by a clean standalone run (with "mvn *clean* test"). So now I'm more convinced it's some kind of environment chaos caused by parallel execution of test cases, and not a blocker issue.
@Andrew It seems to me that kerby jar is not included in our binary package, so I'm not sure whether a new RC is required by HBASE-22219. Anyway I'd like to revoke my -1 vote now. Thanks. Best Regards, Yu On Tue, 16 Apr 2019 at 10:19, Yu Li <car...@gmail.com> wrote: > Sorry for the late response due to job priority. > > This ConnectionRefused issue cannot be reproduced on my laptop (MacOS > 10.14.4) but could on the linux env. And I've checked and confirmed it > could pass with 1.4.7/1.4.9 source package but stably failed with 1.5.0, > performing a git bisect now, will report back later. > > Best Regards, > Yu > > > On Sat, 13 Apr 2019 at 00:38, Andrew Purtell <andrew.purt...@gmail.com> > wrote: > >> I also see the occasional ConnectionRefused errors. They don’t reproduce >> if you run the test standalone. I also only see them on a Linux dev host. >> That may be enough to find by bisect the commit that introduced this >> behavior. Working on it. There is a JIRA filed for this one. Search for >> “TestBlocksRead” and label “branch-1”. >> >> Thanks for the investigations. >> >> > On Apr 12, 2019, at 6:36 AM, Yu Li <car...@gmail.com> wrote: >> > >> > Quick updates: >> > >> > W/ patch of HBASE-22219 or say upgrading kerby version to 1.0.1, the >> > failures listed above in the 1st part of hbase-server disappeared. >> > >> > However, in the 2nd part of hbase-server UT there're still many >> > ConnectionRefused exceptions (17 errors in total) as shown below, which >> > could be reproduced easily with -Dtest=xxx command on my environments, >> > still checking the root cause. >> > >> > [INFO] Running org.apache.hadoop.hbase.regionserver.TestBlocksRead >> > [ERROR] Tests run: 4, Failures: 0, Errors: 4, Skipped: 0, Time elapsed: >> > 0.853 s <<< FAILURE! - in >> > org.apache.hadoop.hbase.regionserver.TestBlocksRead >> > [ERROR] >> > >> testBlocksStoredWhenCachingDisabled(org.apache.hadoop.hbase.regionserver.TestBlocksRead) >> > Time elapsed: 0.17 s <<< ERROR! >> > java.net.ConnectException: Call From >> > z05f06378.sqa.zth.tbsite.net/11.163.183.195 to localhost:35669 failed >> on >> > connection exception: java.net.ConnectException: Connection refused; For >> > more details see: >> > http://wiki.apache.org/hadoop/ConnectionRefused >> > at >> > >> org.apache.hadoop.hbase.regionserver.TestBlocksRead.initHRegion(TestBlocksRead.java:112) >> > at >> > >> org.apache.hadoop.hbase.regionserver.TestBlocksRead.testBlocksStoredWhenCachingDisabled(TestBlocksRead.java:389) >> > Caused by: java.net.ConnectException: Connection refused >> > at >> > >> org.apache.hadoop.hbase.regionserver.TestBlocksRead.initHRegion(TestBlocksRead.java:112) >> > at >> > >> org.apache.hadoop.hbase.regionserver.TestBlocksRead.testBlocksStoredWhenCachingDisabled(TestBlocksRead.java:389) >> > >> > Best Regards, >> > Yu >> > >> > >> >> On Fri, 12 Apr 2019 at 13:11, Yu Li <car...@gmail.com> wrote: >> >> >> >> I have no doubt that you've run the tests locally before announcing a >> >> release as you're always a great RM boss. And this shows one value of >> >> verifying release, that different voter has different environments. >> >> >> >> Now I think the failures may be kerberos related, since I possibly has >> >> changed some system configuration when doing Flink testing on this env >> >> weeks ago. Located one issue (HBASE-22219) which also observed in >> 1.4.7, >> >> will further investigate. >> >> >> >> Best Regards, >> >> Yu >> >> >> >> >> >> On Fri, 12 Apr 2019 at 12:38, Andrew Purtell <andrew.purt...@gmail.com >> > >> >> wrote: >> >> >> >>> “However it's good to find the issue earlier if there >> >>> really is any, before release announced.” >> >>> >> >>> I run the complete unit test suite before announcing a release >> candidate. >> >>> Just to be clear. >> >>> >> >>> Totally agree we should get these problems sorted before an actual >> >>> release. My policy is to cancel a RC if anyone vetoes for this >> reason... >> >>> want as much coverage and varying environments as we can manage. >> >>> >> >>> Thank you for your help so far and I hope the failures you see result >> in >> >>> analysis and fixes that lead to better test stability. >> >>> >> >>>> On Apr 11, 2019, at 9:32 PM, Yu Li <car...@gmail.com> wrote: >> >>>> >> >>>> Confirmed in 1.4.7 source the listed out cases passed (all in the 1st >> >>> part >> >>>> of hbase-server so the result comes out quickly.)... Also confirmed >> the >> >>>> test ran order are the same... >> >>>> >> >>>> Will try 1.5.0 again to prevent the environment difference caused by >> >>> time. >> >>>> If 1.5.0 still fails, will start to do the git bisect to locate the >> >>> first >> >>>> bad commit. >> >>>> >> >>>> Was also expecting an easy pass and +1 as always to save time and >> >>> efforts, >> >>>> but obvious no luck. However it's good to find the issue earlier if >> >>> there >> >>>> really is any, before release announced. >> >>>> >> >>>> Best Regards, >> >>>> Yu >> >>>> >> >>>> >> >>>>> On Fri, 12 Apr 2019 at 12:16, Yu Li <car...@gmail.com> wrote: >> >>>>> >> >>>>> Fine, let's focus on verifying whether it's a real problem rather >> than >> >>>>> arguing about wording, after all that's not my intention... >> >>>>> >> >>>>> As mentioned, I participated in the 1.4.7 release vote[1] and IIRC I >> >>> was >> >>>>> using the same env and all tests passed w/o issue, that's where my >> >>> concern >> >>>>> lies and the main reason I gave a -1 vote. I'm running against 1.4.7 >> >>> source >> >>>>> on the same now and let's see the result. >> >>>>> >> >>>>> [1] https://www.mail-archive.com/dev@hbase.apache.org/msg51380.html >> >>>>> >> >>>>> Best Regards, >> >>>>> Yu >> >>>>> >> >>>>> >> >>>>> On Fri, 12 Apr 2019 at 12:05, Andrew Purtell < >> andrew.purt...@gmail.com >> >>>> >> >>>>> wrote: >> >>>>> >> >>>>>> I believe the test execution order matters. We run some tests in >> >>>>>> parallel. The ordering of tests is determined by readdir() results >> >>> and this >> >>>>>> differs from host to host and checkout to checkout. So when you >> see a >> >>>>>> repeatable group of failures, that’s great. And when someone else >> >>> doesn’t >> >>>>>> see those same tests fail, or they cannot be reproduced when >> running >> >>> by >> >>>>>> themselves, the commonly accepted term of art for this is “flaky”. >> >>>>>> >> >>>>>> >> >>>>>>> On Apr 11, 2019, at 8:52 PM, Yu Li <car...@gmail.com> wrote: >> >>>>>>> >> >>>>>>> Sorry but I'd call it "possible environment related problem" or >> "some >> >>>>>>> feature may not work well in specific environment", rather than a >> >>> flaky. >> >>>>>>> >> >>>>>>> Will check against 1.4.7 released source package before opening >> any >> >>>>>> JIRA. >> >>>>>>> >> >>>>>>> Best Regards, >> >>>>>>> Yu >> >>>>>>> >> >>>>>>> >> >>>>>>> On Fri, 12 Apr 2019 at 11:37, Andrew Purtell < >> >>> andrew.purt...@gmail.com> >> >>>>>>> wrote: >> >>>>>>> >> >>>>>>>> And if they pass in my environment , then what should we call it >> >>> then. >> >>>>>> I >> >>>>>>>> have no doubt you are seeing failures. Therefore can you please >> file >> >>>>>> JIRAs >> >>>>>>>> and attach information that can help identify a fix. Thanks. >> >>>>>>>> >> >>>>>>>>> On Apr 11, 2019, at 8:35 PM, Yu Li <car...@gmail.com> wrote: >> >>>>>>>>> >> >>>>>>>>> I ran the test suite with the >> -Dsurefire.rerunFailingTestsCount=2 >> >>>>>> option >> >>>>>>>>> and on two different env separately, so it sums up to 6 times >> >>> stable >> >>>>>>>>> failure for each case, and from my perspective this is not >> flaky. >> >>>>>>>>> >> >>>>>>>>> IIRC last time when verifying 1.4.7 on the same env no such >> issue >> >>>>>>>> observed, >> >>>>>>>>> will double check. >> >>>>>>>>> >> >>>>>>>>> Best Regards, >> >>>>>>>>> Yu >> >>>>>>>>> >> >>>>>>>>> >> >>>>>>>>> On Fri, 12 Apr 2019 at 00:07, Andrew Purtell < >> >>>>>> andrew.purt...@gmail.com> >> >>>>>>>>> wrote: >> >>>>>>>>> >> >>>>>>>>>> There are two failure cases it looks like. And this looks like >> >>>>>> flakes. >> >>>>>>>>>> >> >>>>>>>>>> The wrong FS assertions are not something I see when I run >> these >> >>>>>> tests >> >>>>>>>>>> myself. I am not able to investigate something I can’t >> reproduce. >> >>>>>> What I >> >>>>>>>>>> suggest is since you can reproduce do a git bisect to find the >> >>> commit >> >>>>>>>> that >> >>>>>>>>>> introduced the problem. Then we can revert it. As an >> alternative >> >>> we >> >>>>>> can >> >>>>>>>>>> open a JIRA, report the problem, temporarily @ignore the test, >> and >> >>>>>>>>>> continue. This latter option only should be done if we are >> fairly >> >>>>>>>> confident >> >>>>>>>>>> it is a test only problem. >> >>>>>>>>>> >> >>>>>>>>>> The connect exceptions are interesting. I see these sometimes >> when >> >>>>>> the >> >>>>>>>>>> suite is executed, not this particular case, but when the >> failed >> >>>>>> test is >> >>>>>>>>>> executed by itself it always passes. It is possible some >> change to >> >>>>>>>> classes >> >>>>>>>>>> related to the minicluster or startup or shutdown timing are >> the >> >>>>>> cause, >> >>>>>>>> but >> >>>>>>>>>> it is test time flaky behavior. I’m not happy about this but it >> >>>>>> doesn’t >> >>>>>>>>>> actually fail the release because the failure is never >> repeatable >> >>>>>> when >> >>>>>>>> the >> >>>>>>>>>> test is run standalone. >> >>>>>>>>>> >> >>>>>>>>>> In general it would be great if some attention was paid to test >> >>>>>>>>>> cleanliness on branch-1. As RM I’m not in a position to insist >> >>> that >> >>>>>>>>>> everything is perfect or there will never be another 1.x >> release, >> >>>>>>>> certainly >> >>>>>>>>>> not from branch-1. So, tests which fail repeatedly block a >> release >> >>>>>> IMHO >> >>>>>>>> but >> >>>>>>>>>> flakes do not. >> >>>>>>>>>> >> >>>>>>>>>> >> >>>>>>>>>>> On Apr 10, 2019, at 11:20 PM, Yu Li <car...@gmail.com> wrote: >> >>>>>>>>>>> >> >>>>>>>>>>> -1 >> >>>>>>>>>>> >> >>>>>>>>>>> Observed many UT failures when checking the source package >> (tried >> >>>>>>>>>> multiple >> >>>>>>>>>>> rounds on two different environments, MacOs and Linux, got the >> >>> same >> >>>>>>>>>>> result), including (but not limited to): >> >>>>>>>>>>> >> >>>>>>>>>>> TestBulkload: >> >>>>>>>>>>> >> >>>>>>>>>> >> >>>>>>>> >> >>>>>> >> >>> >> shouldBulkLoadSingleFamilyHLog(org.apache.hadoop.hbase.regionserver.TestBulkLoad) >> >>>>>>>>>>> Time elapsed: 0.083 s <<< ERROR! >> >>>>>>>>>>> java.lang.IllegalArgumentException: Wrong FS: >> >>>>>>>>>>> >> >>>>>>>>>> >> >>>>>>>> >> >>>>>> >> >>> >> file:/var/folders/t6/vch4nh357f98y1wlq09lbm7h0000gn/T/junit1805329913454564189/junit8020757893576011944/data/default/shouldBulkLoadSingleFamilyHLog/8f4a6b584533de2fd1bf3c398dfaac29, >> >>>>>>>>>>> expected: hdfs://localhost:55938 >> >>>>>>>>>>> at >> >>>>>>>>>>> >> >>>>>>>>>> >> >>>>>>>> >> >>>>>> >> >>> >> org.apache.hadoop.hbase.regionserver.TestBulkLoad.testRegionWithFamiliesAndSpecifiedTableName(TestBulkLoad.java:246) >> >>>>>>>>>>> at >> >>>>>>>>>>> >> >>>>>>>>>> >> >>>>>>>> >> >>>>>> >> >>> >> org.apache.hadoop.hbase.regionserver.TestBulkLoad.testRegionWithFamilies(TestBulkLoad.java:256) >> >>>>>>>>>>> at >> >>>>>>>>>>> >> >>>>>>>>>> >> >>>>>>>> >> >>>>>> >> >>> >> org.apache.hadoop.hbase.regionserver.TestBulkLoad.shouldBulkLoadSingleFamilyHLog(TestBulkLoad.java:150) >> >>>>>>>>>>> >> >>>>>>>>>>> TestStoreFile: >> >>>>>>>>>>> >> >>>>>>>>>> >> >>>>>>>> >> >>>>>> >> >>> >> testCacheOnWriteEvictOnClose(org.apache.hadoop.hbase.regionserver.TestStoreFile) >> >>>>>>>>>>> Time elapsed: 0.083 s <<< ERROR! >> >>>>>>>>>>> java.net.ConnectException: Call From localhost/127.0.0.1 to >> >>>>>>>>>> localhost:55938 >> >>>>>>>>>>> failed on connection exception: java.net.ConnectException: >> >>>>>> Connection >> >>>>>>>>>>> refused; For more details see: >> >>>>>>>>>>> http://wiki.apache.org/hadoop/ConnectionRefused >> >>>>>>>>>>> at >> >>>>>>>>>>> >> >>>>>>>>>> >> >>>>>>>> >> >>>>>> >> >>> >> org.apache.hadoop.hbase.regionserver.TestStoreFile.writeStoreFile(TestStoreFile.java:1047) >> >>>>>>>>>>> at >> >>>>>>>>>>> >> >>>>>>>>>> >> >>>>>>>> >> >>>>>> >> >>> >> org.apache.hadoop.hbase.regionserver.TestStoreFile.testCacheOnWriteEvictOnClose(TestStoreFile.java:908) >> >>>>>>>>>>> >> >>>>>>>>>>> TestHFile: >> >>>>>>>>>>> testEmptyHFile(org.apache.hadoop.hbase.io.hfile.TestHFile) >> Time >> >>>>>>>> elapsed: >> >>>>>>>>>>> 0.08 s <<< ERROR! >> >>>>>>>>>>> java.net.ConnectException: Call From >> >>>>>>>>>>> z05f06378.sqa.zth.tbsite.net/11.163.183.195 to >> localhost:35529 >> >>>>>> failed >> >>>>>>>> on >> >>>>>>>>>>> connection exception: java.net.ConnectException: Connection >> >>> refused; >> >>>>>>>> For >> >>>>>>>>>>> more details see: >> >>> http://wiki.apache.org/hadoop/ConnectionRefused >> >>>>>>>>>>> at >> >>>>>>>>>>> org.apache.hadoop.hbase.io >> >>>>>>>>>> .hfile.TestHFile.testEmptyHFile(TestHFile.java:90) >> >>>>>>>>>>> Caused by: java.net.ConnectException: Connection refused >> >>>>>>>>>>> at >> >>>>>>>>>>> org.apache.hadoop.hbase.io >> >>>>>>>>>> .hfile.TestHFile.testEmptyHFile(TestHFile.java:90) >> >>>>>>>>>>> >> >>>>>>>>>>> TestBlocksScanned: >> >>>>>>>>>>> >> >>>>>>>>>> >> >>>>>>>> >> >>>>>> >> >>> >> testBlocksScannedWithEncoding(org.apache.hadoop.hbase.regionserver.TestBlocksScanned) >> >>>>>>>>>>> Time elapsed: 0.069 s <<< ERROR! >> >>>>>>>>>>> java.lang.IllegalArgumentException: Wrong FS: >> >>>>>>>> hdfs://localhost:35529/tmp/ >> >>>>>>>>>>> >> >>>>>>>>>> >> >>>>>>>> >> >>>>>> >> >>> >> hbase-jueding.ly/hbase/data/default/TestBlocksScannedWithEncoding/a4a416cc3060d9820a621c294af0aa08 >> >>>>>>>>>> , >> >>>>>>>>>>> expected: file:/// >> >>>>>>>>>>> at >> >>>>>>>>>>> >> >>>>>>>>>> >> >>>>>>>> >> >>>>>> >> >>> >> org.apache.hadoop.hbase.regionserver.TestBlocksScanned._testBlocksScanned(TestBlocksScanned.java:90) >> >>>>>>>>>>> at >> >>>>>>>>>>> >> >>>>>>>>>> >> >>>>>>>> >> >>>>>> >> >>> >> org.apache.hadoop.hbase.regionserver.TestBlocksScanned.testBlocksScannedWithEncoding(TestBlocksScanned.java:86) >> >>>>>>>>>>> >> >>>>>>>>>>> And please let me know if any known issue I'm not aware of. >> >>> Thanks. >> >>>>>>>>>>> >> >>>>>>>>>>> Best Regards, >> >>>>>>>>>>> Yu >> >>>>>>>>>>> >> >>>>>>>>>>> >> >>>>>>>>>>>> On Mon, 8 Apr 2019 at 11:38, Yu Li <car...@gmail.com> wrote: >> >>>>>>>>>>>> >> >>>>>>>>>>>> The performance report LGTM, thanks! (and sorry for the lag >> due >> >>> to >> >>>>>>>>>>>> Qingming Festival Holiday here in China) >> >>>>>>>>>>>> >> >>>>>>>>>>>> Still verifying the release, just some quick feedback: >> observed >> >>>>>> some >> >>>>>>>>>>>> incompatible changes in compatibility report including >> >>>>>>>>>>>> HBASE-21492/HBASE-21684 and worth a reminder in ReleaseNote. >> >>>>>>>>>>>> >> >>>>>>>>>>>> Irrelative but noticeable: the 1.4.9 release note URL is >> >>> invalid on >> >>>>>>>>>>>> https://hbase.apache.org/downloads.html >> >>>>>>>>>>>> >> >>>>>>>>>>>> Best Regards, >> >>>>>>>>>>>> Yu >> >>>>>>>>>>>> >> >>>>>>>>>>>> >> >>>>>>>>>>>>> On Fri, 5 Apr 2019 at 08:45, Andrew Purtell < >> >>> apurt...@apache.org> >> >>>>>>>>>> wrote: >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> The difference is basically noise per the usual YCSB >> >>> evaluation. >> >>>>>>>> Small >> >>>>>>>>>>>>> differences in workloads D and F (slightly worse) and >> workload >> >>> E >> >>>>>>>>>> (slightly >> >>>>>>>>>>>>> better) that do not indicate serious regression. >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> Linux version 4.14.55-62.37.amzn1.x86_64 >> >>>>>>>>>>>>> c3.8xlarge x 5 >> >>>>>>>>>>>>> OpenJDK Runtime Environment (build 1.8.0_181-shenandoah-b13) >> >>>>>>>>>>>>> -Xms20g -Xmx20g -XX:+UseG1GC -XX:+AlwaysPreTouch >> -XX:+UseNUMA >> >>>>>>>>>>>>> -XX:-UseBiasedLocking -XX:+ParallelRefProcEnabled >> >>>>>>>>>>>>> Hadoop 2.9.2 >> >>>>>>>>>>>>> Init: Load 100 M rows and snapshot >> >>>>>>>>>>>>> Run: Delete table, clone and redeploy from snapshot, run 10 >> M >> >>>>>>>>>> operations >> >>>>>>>>>>>>> Args: -threads 100 -target 50000 >> >>>>>>>>>>>>> Test table: {NAME => 'u', BLOOMFILTER => 'ROW', VERSIONS => >> >>> '1', >> >>>>>>>>>> IN_MEMORY >> >>>>>>>>>>>>> => 'false', KEEP_DELETED_CELLS => 'FALSE', >> DATA_BLOCK_ENCODING >> >>> => >> >>>>>>>>>>>>> 'ROW_INDEX_V1', TTL => 'FOREVER', COMPRESSION => 'SNAPPY', >> >>>>>>>>>> MIN_VERSIONS => >> >>>>>>>>>>>>> '0', BLOCKCACHE => 'true', BLOCKSIZE => '65536', >> >>>>>> REPLICATION_SCOPE => >> >>>>>>>>>>>>> '0'} >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> YCSB Workload A >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> target 50k/op/s 1.4.9 1.5.0 >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> [OVERALL], RunTime(ms) 200592 200583 >> >>>>>>>>>>>>> [OVERALL], Throughput(ops/sec) 49852 49855 >> >>>>>>>>>>>>> [READ], AverageLatency(us) 544 559 >> >>>>>>>>>>>>> [READ], MinLatency(us) 267 292 >> >>>>>>>>>>>>> [READ], MaxLatency(us) 165631 185087 >> >>>>>>>>>>>>> [READ], 95thPercentileLatency(us) 738 742 >> >>>>>>>>>>>>> [READ], 99thPercentileLatency(us), 1877 1961 >> >>>>>>>>>>>>> [UPDATE], AverageLatency(us) 1370 1181 >> >>>>>>>>>>>>> [UPDATE], MinLatency(us) 702 646 >> >>>>>>>>>>>>> [UPDATE], MaxLatency(us) 180735 177279 >> >>>>>>>>>>>>> [UPDATE], 95thPercentileLatency(us) 1943 1652 >> >>>>>>>>>>>>> [UPDATE], 99thPercentileLatency(us) 3257 3085 >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> YCSB Workload B >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> target 50k/op/s 1.4.9 1.5.0 >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> [OVERALL], RunTime(ms) 200599 200581 >> >>>>>>>>>>>>> [OVERALL], Throughput(ops/sec) 49850 49855 >> >>>>>>>>>>>>> [READ], AverageLatency(us), 454 471 >> >>>>>>>>>>>>> [READ], MinLatency(us) 203 213 >> >>>>>>>>>>>>> [READ], MaxLatency(us) 183423 174207 >> >>>>>>>>>>>>> [READ], 95thPercentileLatency(us) 563 599 >> >>>>>>>>>>>>> [READ], 99thPercentileLatency(us) 1360 1172 >> >>>>>>>>>>>>> [UPDATE], AverageLatency(us) 1064 1029 >> >>>>>>>>>>>>> [UPDATE], MinLatency(us) 746 726 >> >>>>>>>>>>>>> [UPDATE], MaxLatency(us) 163455 101631 >> >>>>>>>>>>>>> [UPDATE], 95thPercentileLatency(us) 1327 1157 >> >>>>>>>>>>>>> [UPDATE], 99thPercentileLatency(us) 2241 1898 >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> YCSB Workload C >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> target 50k/op/s 1.4.9 1.5.0 >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> [OVERALL], RunTime(ms) 200541 200538 >> >>>>>>>>>>>>> [OVERALL], Throughput(ops/sec) 49865 49865 >> >>>>>>>>>>>>> [READ], AverageLatency(us) 332 327 >> >>>>>>>>>>>>> [READ], MinLatency(us) 175 179 >> >>>>>>>>>>>>> [READ], MaxLatency(us) 210559 170367 >> >>>>>>>>>>>>> [READ], 95thPercentileLatency(us) 410 396 >> >>>>>>>>>>>>> [READ], 99thPercentileLatency(us) 871 892 >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> YCSB Workload D >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> target 50k/op/s 1.4.9 1.5.0 >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> [OVERALL], RunTime(ms) 200579 200562 >> >>>>>>>>>>>>> [OVERALL], Throughput(ops/sec) 49855 49859 >> >>>>>>>>>>>>> [READ], AverageLatency(us) 487 547 >> >>>>>>>>>>>>> [READ], MinLatency(us) 210 214 >> >>>>>>>>>>>>> [READ], MaxLatency(us) 192255 177535 >> >>>>>>>>>>>>> [READ], 95thPercentileLatency(us) 973 1529 >> >>>>>>>>>>>>> [READ], 99thPercentileLatency(us) 1836 2683 >> >>>>>>>>>>>>> [INSERT], AverageLatency(us) 1239 1152 >> >>>>>>>>>>>>> [INSERT], MinLatency(us) 807 788 >> >>>>>>>>>>>>> [INSERT], MaxLatency(us) 184575 148735 >> >>>>>>>>>>>>> [INSERT], 95thPercentileLatency(us) 1496 1243 >> >>>>>>>>>>>>> [INSERT], 99thPercentileLatency(us) 2965 2495 >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> YCSB Workload E >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> target 10k/op/s 1.4.9 1.5.0 >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> [OVERALL], RunTime(ms) 100605 100568 >> >>>>>>>>>>>>> [OVERALL], Throughput(ops/sec) 9939 9943 >> >>>>>>>>>>>>> [SCAN], AverageLatency(us) 3548 2687 >> >>>>>>>>>>>>> [SCAN], MinLatency(us) 696 678 >> >>>>>>>>>>>>> [SCAN], MaxLatency(us) 1059839 238463 >> >>>>>>>>>>>>> [SCAN], 95thPercentileLatency(us) 8327 6791 >> >>>>>>>>>>>>> [SCAN], 99thPercentileLatency(us) 17647 14415 >> >>>>>>>>>>>>> [INSERT], AverageLatency(us) 2688 1555 >> >>>>>>>>>>>>> [INSERT], MinLatency(us) 887 815 >> >>>>>>>>>>>>> [INSERT], MaxLatency(us) 173311 154623 >> >>>>>>>>>>>>> [INSERT], 95thPercentileLatency(us) 4455 2571 >> >>>>>>>>>>>>> [INSERT], 99thPercentileLatency(us) 9303 5375 >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> YCSB Workload F >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> target 50k/op/s 1.4.9 1.5.0 >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> [OVERALL], RunTime(ms) 200562 204178 >> >>>>>>>>>>>>> [OVERALL], Throughput(ops/sec) 49859 48976 >> >>>>>>>>>>>>> [READ], AverageLatency(us) 856 1137 >> >>>>>>>>>>>>> [READ], MinLatency(us) 262 257 >> >>>>>>>>>>>>> [READ], MaxLatency(us) 205567 222335 >> >>>>>>>>>>>>> [READ], 95thPercentileLatency(us) 2365 3475 >> >>>>>>>>>>>>> [READ], 99thPercentileLatency(us) 3099 4143 >> >>>>>>>>>>>>> [READ-MODIFY-WRITE], AverageLatency(us) 2559 2917 >> >>>>>>>>>>>>> [READ-MODIFY-WRITE], MinLatency(us) 1100 1034 >> >>>>>>>>>>>>> [READ-MODIFY-WRITE], MaxLatency(us) 208767 204799 >> >>>>>>>>>>>>> [READ-MODIFY-WRITE], 95thPercentileLatency(us) 5747 7627 >> >>>>>>>>>>>>> [READ-MODIFY-WRITE], 99thPercentileLatency(us) 7203 8919 >> >>>>>>>>>>>>> [UPDATE], AverageLatency(us) 1700 1777 >> >>>>>>>>>>>>> [UPDATE], MinLatency(us) 737 687 >> >>>>>>>>>>>>> [UPDATE], MaxLatency(us) 97983 94271 >> >>>>>>>>>>>>> [UPDATE], 95thPercentileLatency(us) 3377 4147 >> >>>>>>>>>>>>> [UPDATE], 99thPercentileLatency(us) 4147 4831 >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>>> On Thu, Apr 4, 2019 at 1:14 AM Yu Li <car...@gmail.com> >> >>> wrote: >> >>>>>>>>>>>>>> >> >>>>>>>>>>>>>> Thanks for the efforts boss. >> >>>>>>>>>>>>>> >> >>>>>>>>>>>>>> Since it's a new minor release, do we have performance >> >>> comparison >> >>>>>>>>>> report >> >>>>>>>>>>>>>> with 1.4.9 as we did when releasing 1.4.0? If so, any >> >>> reference? >> >>>>>>>> Many >> >>>>>>>>>>>>>> thanks! >> >>>>>>>>>>>>>> >> >>>>>>>>>>>>>> Best Regards, >> >>>>>>>>>>>>>> Yu >> >>>>>>>>>>>>>> >> >>>>>>>>>>>>>> >> >>>>>>>>>>>>>> On Thu, 4 Apr 2019 at 07:44, Andrew Purtell < >> >>> apurt...@apache.org >> >>>>>>> >> >>>>>>>>>>>>> wrote: >> >>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> The fourth HBase 1.5.0 release candidate (RC3) is >> available >> >>> for >> >>>>>>>>>>>>> download >> >>>>>>>>>>>>>> at >> >>>>>>>>>>>>>>> >> https://dist.apache.org/repos/dist/dev/hbase/hbase-1.5.0RC3/ >> >>>>>> and >> >>>>>>>>>>>>> Maven >> >>>>>>>>>>>>>>> artifacts are available in the temporary repository >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>> >> >>>>>> >> >>> >> https://repository.apache.org/content/repositories/orgapachehbase-1292/ >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> The git tag corresponding to the candidate is '1.5.0RC3’ >> >>>>>>>>>> (b0bc7225c5). >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> A detailed source and binary compatibility report for this >> >>>>>> release >> >>>>>>>> is >> >>>>>>>>>>>>>>> available for your review at >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>> >> >>>>>>>> >> >>>>>> >> >>> >> https://dist.apache.org/repos/dist/dev/hbase/hbase-1.5.0RC3/compat-check-report.html >> >>>>>>>>>>>>>>> . >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> A list of the 115 issues resolved in this release can be >> >>> found >> >>>>>> at >> >>>>>>>>>>>>>>> https://s.apache.org/K4Wk . The 1.5.0 changelog is >> derived >> >>> from >> >>>>>>>> the >> >>>>>>>>>>>>>>> changelog of the last branch-1.4 release, 1.4.9. >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> Please try out the candidate and vote +1/0/-1. >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> The vote will be open for at least 72 hours. Unless >> >>> objection I >> >>>>>>>> will >> >>>>>>>>>>>>> try >> >>>>>>>>>>>>>> to >> >>>>>>>>>>>>>>> close it Friday April 12, 2019 if we have sufficient >> votes. >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> Prior to making this announcement I made the following >> >>> preflight >> >>>>>>>>>>>>> checks: >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> RAT check passes (7u80) >> >>>>>>>>>>>>>>> Unit test suite passes (7u80, 8u181)* >> >>>>>>>>>>>>>>> Opened the UI in a browser, poked around >> >>>>>>>>>>>>>>> LTT load 100M rows with 100% verification and 20% updates >> >>>>>> (8u181) >> >>>>>>>>>>>>>>> ITBLL 1B rows with slowDeterministic monkey (8u181) >> >>>>>>>>>>>>>>> ITBLL 1B rows with serverKilling monkey (8u181) >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> There are known flaky tests. See HBASE-21904 and >> HBASE-21905. >> >>>>>> These >> >>>>>>>>>>>>> flaky >> >>>>>>>>>>>>>>> tests do not represent serious test failures that would >> >>> prevent >> >>>>>> a >> >>>>>>>>>>>>>> release. >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>>> -- >> >>>>>>>>>>>>>>> Best regards, >> >>>>>>>>>>>>>>> Andrew >> >>>>>>>>>>>>>>> >> >>>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> -- >> >>>>>>>>>>>>> Best regards, >> >>>>>>>>>>>>> Andrew >> >>>>>>>>>>>>> >> >>>>>>>>>>>>> Words like orphans lost among the crosstalk, meaning torn >> from >> >>>>>>>> truth's >> >>>>>>>>>>>>> decrepit hands >> >>>>>>>>>>>>> - A23, Crosstalk >> >>>>>>>>>>>>> >> >>>>>>>>>>>> >> >>>>>>>>>> >> >>>>>>>> >> >>>>>> >> >>>>> >> >>> >> >> >> >