Hi Andy,

This email must have caught attention of a number of people...
You mention "Linux AMI (2012.03.1)", but which AMI is that?  Is this some 
specific AMI prepared by Amazon?  Or some AMI that somebody like Cloudera 
prepared?  Or are you saying it's just "some Linux" AMI that somebody built on 
2012-03-01 and that you found in AWS?

Could you please share the outputs of:

$ cat /etc/*release
$ uname -a

$ df -T

Also, could it be that your old EC2 instance was unlucky and had a very noisy 
neighbour, while the new EC2 instance does not?  Not sure how one could run 
tests to get around this - perhaps by terminating the instance and restarting 
it a few times in order to get it hosted on different physical hosts?

Thanks,
Otis 
----
Performance Monitoring SaaS for HBase - 
http://sematext.com/spm/hbase-performance-monitoring/index.html



>________________________________
> From: Andrew Purtell <[email protected]>
>To: "[email protected]" <[email protected]> 
>Cc: Jack Levin <[email protected]>; "[email protected]" 
><[email protected]> 
>Sent: Tuesday, April 10, 2012 2:14 PM
>Subject: Re: Speeding up HBase read response
> 
>What AMI are you using as your base?
>
>I recently started using the new Linux AMI (2012.03.1) and noticed what looks 
>like significant improvement over what I had been using before (2011.02 IIRC). 
>I ran four simple tests repeated three times with FIO: a read bandwidth test, 
>a write bandwidth test, a read IOPS test, and a write IOPS test. The write 
>IOPS test was inconclusive but for the others there was a consistent 
>difference: reduced disk op latency (shorter tail) and increased device 
>bandwidth. I don't run anything in production in EC2 so this was the extent of 
>my curiosity.
>
>
>Best regards,
>
>    - Andy
>
>Problems worthy of attack prove their worth by hitting back. - Piet Hein (via 
>Tom White)
>
>
>
>----- Original Message -----
>> From: Jeff Whiting <[email protected]>
>> To: [email protected]
>> Cc: Jack Levin <[email protected]>; [email protected]
>> Sent: Tuesday, April 10, 2012 11:03 AM
>> Subject: Re: Speeding up HBase read response
>> 
>> Do you have bloom filters enabled?  And compression?  Both of those can help 
>> reduce disk io load 
>> which seems to be the main issue you are having on the ec2 cluster.
>> 
>> ~Jeff
>> 
>> On 4/9/2012 8:28 AM, Jack Levin wrote:
>>>  Yes, from  %util you can see that your disks are working at 100%
>>>  pretty much.  Which means you can't push them go any faster.   So the
>>>  solution is to add more disks, add faster disks, add nodes and disks.
>>>  This type of overload should not be related to HBASE, but rather to
>>>  your hardware setup.
>>> 
>>>  -Jack
>>> 
>>>  On Mon, Apr 9, 2012 at 2:29 AM, ijanitran<[email protected]>  wrote:
>>>>  Hi, results of iostat are pretty much very similar on all nodes:
>>>> 
>>>>  Device:         rrqm/s   wrqm/s     r/s     w/s    rMB/s    wMB/s 
>> avgrq-sz
>>>>  avgqu-sz   await  svctm  %util
>>>>  xvdap1            0.00     0.00  294.00    0.00     9.27     0.00    
>> 64.54
>>>>  21.97   75.44   3.40 100.10
>>>> 
>>>>  Device:         rrqm/s   wrqm/s     r/s     w/s    rMB/s    wMB/s 
>> avgrq-sz
>>>>  avgqu-sz   await  svctm  %util
>>>>  xvdap1            0.00     4.00  286.00    8.00     9.11     0.27    
>> 65.33
>>>>  7.16 25.32 2.88  84.70
>>>> 
>>>>  Device:         rrqm/s   wrqm/s     r/s     w/s    rMB/s    wMB/s 
>> avgrq-sz
>>>>  avgqu-sz   await  svctm  %util
>>>>  xvdap1            0.00     0.00  283.00    0.00     8.29     0.00    
>> 59.99
>>>>  10.31   35.43   2.97  84.10
>>>> 
>>>>  Device:         rrqm/s   wrqm/s     r/s     w/s    rMB/s    wMB/s 
>> avgrq-sz
>>>>  avgqu-sz   await  svctm  %util
>>>>  xvdap1            0.00     0.00  320.00    0.00     9.12     0.00    
>> 58.38
>>>>  12.32   39.56   2.79  89.40
>>>> 
>>>>  Device:         rrqm/s   wrqm/s     r/s     w/s    rMB/s    wMB/s 
>> avgrq-sz
>>>>  avgqu-sz   await  svctm  %util
>>>>  xvdap1            0.00     0.00  336.63    0.00     9.18     0.00    
>> 55.84
>>>>  10.67   31.42   2.78  93.47
>>>> 
>>>>  Device:         rrqm/s   wrqm/s     r/s     w/s    rMB/s    wMB/s 
>> avgrq-sz
>>>>  avgqu-sz   await  svctm  %util
>>>>  xvdap1            0.00     0.00  312.00    0.00    10.00     0.00    
>> 65.62
>>>>  11.07   35.49   2.91  90.70
>>>> 
>>>>  Device:         rrqm/s   wrqm/s     r/s     w/s    rMB/s    wMB/s 
>> avgrq-sz
>>>>  avgqu-sz   await  svctm  %util
>>>>  xvdap1            0.00     0.00  356.00    0.00    10.72     0.00    
>> 61.66
>>>>  9.38 26.63 2.57  91.40
>>>> 
>>>>  Device:         rrqm/s   wrqm/s     r/s     w/s    rMB/s    wMB/s 
>> avgrq-sz
>>>>  avgqu-sz   await  svctm  %util
>>>>  xvdap1            0.00     0.00  258.00    0.00     8.20     0.00    
>> 65.05
>>>>  13.37   51.24   3.64  93.90
>>>> 
>>>>  Device:         rrqm/s   wrqm/s     r/s     w/s    rMB/s    wMB/s 
>> avgrq-sz
>>>>  avgqu-sz   await  svctm  %util
>>>>  xvdap1            0.00     0.00  246.00    0.00     7.31     0.00    
>> 60.88
>>>>  5.87   24.53   3.14  77.30
>>>> 
>>>>  Device:         rrqm/s   wrqm/s     r/s     w/s    rMB/s    wMB/s 
>> avgrq-sz
>>>>  avgqu-sz   await  svctm  %util
>>>>  xvdap1            0.00     2.00  297.00    3.00     9.11     0.02    
>> 62.29
>>>>  13.02   42.40   3.12  93.60
>>>> 
>>>>  Device:         rrqm/s   wrqm/s     r/s     w/s    rMB/s    wMB/s 
>> avgrq-sz
>>>>  avgqu-sz   await  svctm  %util
>>>>  xvdap1            0.00     0.00  292.00    0.00     9.60     0.00    
>> 67.32
>>>>  11.30   39.51   3.36  98.00
>>>> 
>>>>  Device:         rrqm/s   wrqm/s     r/s     w/s    rMB/s    wMB/s 
>> avgrq-sz
>>>>  avgqu-sz   await  svctm  %util
>>>>  xvdap1            0.00     4.00  261.00    8.00     7.84     0.27    
>> 61.74
>>>>  16.07   55.72   3.39  91.30
>>>> 
>>>> 
>>>>  Jack Levin wrote:
>>>>>  Please email iostat -xdm 1, run for one minute during load on each 
>> node
>>>>>  --
>>>>>  Sent from my Android phone with K-9 Mail. Please excuse my brevity.
>>>>> 
>>>>>  ijanitran<[email protected]>  wrote:
>>>>> 
>>>>> 
>>>>>  I have 4 nodes HBase v0.90.4-cdh3u3 cluster deployed on Amazon 
>> XLarge
>>>>>  instances (16Gb RAM, 4 cores CPU) with 8Gb heap -Xmx allocated for 
>> HRegion
>>>>>  servers, 2Gb for datanodes. HMaster\ZK\Namenode is on the 
>> separate XLarge
>>>>>  instance. Target dataset is 100 millions records (each record is 10 
>> fields
>>>>>  by 100 bytes). Benchmarking performed concurrently from parallel 
>> 100
>>>>>  threads.
>>>>> 
>>>>>  I'm confused with a read latency I got, comparing to what YCSB 
>> team
>>>>>  achieved
>>>>>  and showed in their YCSB paper. They achieved throughput of up to 
>> 7000
>>>>>  ops/sec with a latency of 15 ms (page 10, read latency chart). I 
>> can't get
>>>>>  throughput higher than 2000 ops/sec on 90% reads/10% writes 
>> workload.
>>>>>  Writes
>>>>>  are really fast with auto commit disabled (response within a few 
>> ms),
>>>>>  while
>>>>>  read latency doesn't go lower than 70 ms in average.
>>>>> 
>>>>>  These are some HBase settings I used:
>>>>> 
>>>>>  hbase.regionserver.handler.count=50
>>>>>  hfile.block.cache.size=0.4
>>>>>  hbase.hregion.max.filesize=1073741824
>>>>>  hbase.regionserver.codecs=lzo
>>>>>  hbase.hregion.memstore.mslab.enabled=true
>>>>>  hfile.min.blocksize.size=16384
>>>>>  hbase.hregion.memstore.block.multiplier=4
>>>>>  hbase.regionserver.global.memstore.upperLimit=0.35
>>>>>  hbase.zookeeper.property.maxClientCnxns=100
>>>>> 
>>>>>  Which settings do you recommend to look at\tune to speed up 
>> reads with
>>>>>  HBase?
>>>>> 
>>>>>  --
>>>>>  View this message in context:
>>>>> 
>> http://old.nabble.com/Speeding-up-HBase-read-response-tp33635226p33635226.html
>>>>>  Sent from the HBase User mailing list archive at Nabble.com.
>>>>> 
>>>>> 
>>>>> 
>>>>  --
>>>>  View this message in context: 
>> http://old.nabble.com/Speeding-up-HBase-read-response-tp33635226p33654666.html
>>>>  Sent from the HBase User mailing list archive at Nabble.com.
>>>> 
>> 
>> -- 
>> Jeff Whiting
>> Qualtrics Senior Software Engineer
>> [email protected]
>>
>
>
>

Reply via email to