+1, I like that idea.
On Wed, Feb 16, 2011 at 2:43 AM, Todd Lipcon t...@cloudera.com wrote:
Hi Ted,
I'd recommend setting up jdiff to answer this question. Would be a good
contribution to our source base to be able to run this automatically and
generate a report as part of our build. We do
I ran jdiff by hand. See:
https://tm-files.s3.amazonaws.com/hbase/jdiff-hbase-0.90.1/changes.html
Best regards,
- Andy
Problems worthy of attack prove their worth by hitting back.
- Piet Hein (via Tom White)
--- On Wed, 2/16/11, Lars George lars.geo...@gmail.com wrote:
From: Lars
Well done Andrew.
People who want to know the API differences should probably mostly only read:
https://tm-files.s3.amazonaws.com/hbase/jdiff-hbase-0.90.1/changes/pkg_org.apache.hadoop.hbase.client.html
And specifically the HTable, Put, Get, Delete, Scan classes.
On Wed, Feb 16, 2011 at 7:19
Sounds like Ted volunteered to do it!
Good job!
-ryan
On Wed, Feb 16, 2011 at 12:15 PM, Ted Yu yuzhih...@gmail.com wrote:
Definitely.
On Wed, Feb 16, 2011 at 11:57 AM, Todd Lipcon t...@cloudera.com wrote:
In Hadoop land, Tom White did some awesome work to add special annotations
that we
I am not very familiar with (internal) HBase APIs which grow quite large.
I have a full-time job.
And this task is quite big.
Community effort should be the best approach.
On Wed, Feb 16, 2011 at 12:20 PM, Todd Lipcon t...@cloudera.com wrote:
On Wed, Feb 16, 2011 at 12:16 PM, Ryan Rawson
Step 1 is to add the jdiff framework in, that is a non-trivial but
straightforward change.
Step 2 is to annotate all the APIs, something that should be done by
various domain experts over time. Even if this is not complete there
is value with #1.
Step 3: ?
Step 4: profit!
On Wed, Feb 16, 2011 at
+1. On a 5 node test cluster, loaded 100M rows (~170G on disk) using YCSB,
did some mixed workload (increment, scan, read, put). No GC pauses. Killed a
server while under load, recovered correctly.
Along the way I accidentally deleted my ZK jar while the cluster was
running, which left a couple