Ugh, sorry guys, I'm dumb. I was running 1 mapper per RS before, but
switched to a d2.4xlarge instance today and, after noticing cores sitting
idly, decided to try setting the number of mappers and reducers to the
number of cores to speed testing up (RAM is still grossly underutilized
with less than 16 GB/122 GB in use at any one time). This definitely made
runs go faster (generation took less than 3 hours, verification took about
1 hour), but I just realized that the number of nodes I picked (62500000)
isn't a multiple of 25,000,000 and so the list won't wrap properly. I'll
rerun and confirm, but I'm guessing this is a false alarm.

Sorry again. :(

-Dima

On Sun, Sep 4, 2016 at 9:56 PM, Andrew Purtell <andrew.purt...@gmail.com>
wrote:

> I will also try your incantation (and JRE version) on this RC and 0.98.21
> next week to answer those same questions.
>
> Looks like you are using a multiple of RSes (16) as numMappers? Is that
> 4x? On what kind of instance type? I am (also, I think) using a 5 node
> "cluster" with 4 RS nodes but numMappers 4 and numNodes 250000000. Since
> with clusterdock everything is contending for one instance's resources I
> didn't want to overdo and so have started at 1 mapper per RS. Since you
> appear to be using a higher value, I'm curious if you've found that you
> will get stable results with that, if more mappers in this configuration
> does a better job finding problems in your experience, and what instance
> type are you using? I've been using a d2.4xlarge.
>
> > On Sep 4, 2016, at 9:04 PM, Andrew Purtell <andrew.purt...@gmail.com>
> wrote:
> >
> > I've been running 1B tests with slowDeterministic. 0.98.21 and this
> 0.98.22 RC. I get 1B referenced, all ok.
> >
> > Did you run serverKilling with 0.98.21? And did it pass? Or does 0.98.21
> pass for you now? If so then we have a regression. If not then it's
> something to look at for 0.98.23 I'd say.
> >
> >> On Sep 4, 2016, at 8:44 PM, Dima Spivak <dimaspi...@apache.org> wrote:
> >>
> >> Anyone else running ITBLL seeing issues? I just ran a 5-node clusterdock
> >> cluster with JDK 7u79 of this RC and tried out ITBLL with 1 billion rows
> >> and the serverKilling monkey (`hbase
> >> org.apache.hadoop.hbase.test.IntegrationTestBigLinkedList -m
> serverKilling
> >> loop 1 16 62500000 ${RANDOM} 16`). This failed for me because of
> >> unreferenced list nodes:
> >>
> >> org.apache.hadoop.hbase.test.IntegrationTestBigLinkedList$Verify$Counts
> >> REFERENCED=732006926
> >> UNREFERENCED=12003580
> >>
> >> Perhaps this is similar to what Mikhail saw a while back with later
> >> releases?
> >>
> >> -Dima
> >>
> >>> On Sat, Sep 3, 2016 at 8:34 AM, Andrew Purtell <apurt...@apache.org>
> wrote:
> >>>
> >>> The 1st HBase 0.98.2
> >>> ​2 release candidate (RC0) is available for download at
> >>> https://dist.apache.org/repos/dist/dev/hbase/hbase-0.98.22RC0 and
> Maven
> >>> artifacts are also available in the temporary repository
> >>> https://repository.apache.org/content/repositories/orgapachehbase-1151
> .
> >>>
> >>> The detailed source and binary compatibility report for this release
> with
> >>> respect to the previous is available for your review at
> >>> https://dist.apache.org/repos/dist/dev/hbase/hbase-0.98.
> >>> 22RC0/0.98.21_0.98.22RC0_compat_report.html
> >>> ​. There are no reported compatibility issues.
> >>>
> >>> The
> >>> ​25​
> >>> issues resolved in this release can be found at
> https://s.apache.org/C7SV
> >>> .
> >>>
> >>> I have made the following assessments of this candidate:
> >>> - Release audit check
> >>> ​: pass​
> >>>
> >>> -
> >>> ​ Unit test suite: pass 10/10 (7u79)​
> >>>
> >>> - Loaded 1M keys with LTT (10 readers, 10 writers, 10 updaters (20%):
> all
> >>> keys verified, no unusual messages or errors, latencies in the ballpark
> >>> - IntegrationTestBigLinkedList
> >>> ​1B rows: 100% referenced, no errors (8u91)
> >>> - Built head of Apache Phoenix 4.x-HBase-0.98 branch
> >>> ​:​
> >>> ​no errors (7u79)
> >>>
> >>> Signed with my code signing key D5365CCD.
> >>>
> >>> Please try out the candidate and vote +1/0/-1. This vote will be open
> for
> >>> at least 72 hours. Unless objection I will try to close it
> >>> ​Friday September 9, 2016 if we have sufficient votes.
> >>>
> >>> --
> >>> Best regards,
> >>>
> >>>  - Andy
> >>>
> >>> Problems worthy of attack prove their worth by hitting back. - Piet
> Hein
> >>> (via Tom White)
> >>>
>

Reply via email to