The "Thread-Opr0" the client thread to put data into hbase, it is waiting.
"Thread-Opr0-EventThread" daemon prio=10 tid=0x00002aaafc7a8000 nid=0xe08
waiting on condition [0x000000004383f000]
java.lang.Thread.State: WAITING (parking)
at sun.misc.Unsafe.park(Native Method)
- parking to wait for <0x00002aaab632ae50> (a
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
at java.util.concurrent.locks.LockSupport.park(LockSupport.java:158)
at
java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1925)
at
java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399)
at
org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:502)
"Thread-Opr0-SendThread(nd1-rack2-cloud:2181)" daemon prio=10
tid=0x00002aaafc7a6800 nid=0xe07 runnable [0x000000004373e000]
java.lang.Thread.State: RUNNABLE
at sun.nio.ch.EPollArrayWrapper.epollWait(Native Method)
at sun.nio.ch.EPollArrayWrapper.poll(EPollArrayWrapper.java:210)
at sun.nio.ch.EPollSelectorImpl.doSelect(EPollSelectorImpl.java:65)
at sun.nio.ch.SelectorImpl.lockAndDoSelect(SelectorImpl.java:69)
- locked <0x00002aaab6304410> (a sun.nio.ch.Util$1)
- locked <0x00002aaab6304428> (a
java.util.Collections$UnmodifiableSet)
- locked <0x00002aaab632abd0> (a sun.nio.ch.EPollSelectorImpl)
at sun.nio.ch.SelectorImpl.select(SelectorImpl.java:80)
at
org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1107)
"Thread-Opr0" prio=10 tid=0x00002aab0402a000 nid=0xdf2 in Object.wait()
[0x000000004262d000]
java.lang.Thread.State: WAITING (on object monitor)
at java.lang.Object.wait(Native Method)
- waiting on <0x00002aaab04302d0> (a
org.apache.hadoop.hbase.ipc.HBaseClient$Call)
at java.lang.Object.wait(Object.java:485)
at
org.apache.hadoop.hbase.ipc.HBaseClient.call(HBaseClient.java:739)
- locked <0x00002aaab04302d0> (a
org.apache.hadoop.hbase.ipc.HBaseClient$Call)
at
org.apache.hadoop.hbase.ipc.HBaseRPC$Invoker.invoke(HBaseRPC.java:257)
at $Proxy0.getClosestRowBefore(Unknown Source)
at org.apache.hadoop.hbase.client.HTable$3.call(HTable.java:517)
at org.apache.hadoop.hbase.client.HTable$3.call(HTable.java:515)
at
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.getRegionServerWithRetries(HConnectionManager.java:1000)
at
org.apache.hadoop.hbase.client.HTable.getRowOrBefore(HTable.java:514)
at
org.apache.hadoop.hbase.client.MetaScanner.metaScan(MetaScanner.java:133)
at
org.apache.hadoop.hbase.client.MetaScanner.metaScan(MetaScanner.java:95)
at
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.prefetchRegionCache(HConnectionManager.java:645)
at
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegionInMeta(HConnectionManager.java:699)
- locked <0x00002aaab6294660> (a java.lang.Object)
at
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:590)
at
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.processBatch(HConnectionManager.java:1114)
at
org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.processBatchOfPuts(HConnectionManager.java:1234)
at
org.apache.hadoop.hbase.client.HTable.flushCommits(HTable.java:819)
at org.apache.hadoop.hbase.client.HTable.doPut(HTable.java:675)
at org.apache.hadoop.hbase.client.HTable.put(HTable.java:660)
at com.bigdata.bench.hbase.HBaseWriter$Operator.operateTo(Unknown
Source)
at com.bigdata.bench.hbase.HBaseWriter$Operator.run(Unknown Source)
On Thu, Jan 27, 2011 at 12:06 AM, Schubert Zhang <[email protected]> wrote:
> Even though cannot put more data into table, I can read the existing data.
>
> And I stop and re-start the HBase, still cannot put more data.
>
> hbase(main):031:0> status 'simple'
> 8 live servers
> nd5-rack2-cloud:60020 1296057544120
> requests=0, regions=32, usedHeap=130, maxHeap=8973
> nd8-rack2-cloud:60020 1296057544350
> requests=0, regions=31, usedHeap=128, maxHeap=8983
> nd2-rack2-cloud:60020 1296057543346
> requests=0, regions=32, usedHeap=130, maxHeap=8973
> nd3-rack2-cloud:60020 1296057544224
> requests=0, regions=32, usedHeap=133, maxHeap=8973
> nd6-rack2-cloud:60020 1296057544482
> requests=0, regions=32, usedHeap=130, maxHeap=8983
> nd9-rack2-cloud:60020 1296057544565
> requests=174, regions=32, usedHeap=180, maxHeap=8983
> nd7-rack2-cloud:60020 1296057544617
> requests=0, regions=32, usedHeap=126, maxHeap=8983
> nd4-rack2-cloud:60020 1296057544138
> requests=0, regions=32, usedHeap=126, maxHeap=8973
> 0 dead servers
> Aggregate load: 174, regions: 255
>
>
> On Wed, Jan 26, 2011 at 11:58 PM, Schubert Zhang <[email protected]>wrote:
>
>> I am using 0.90.0 (8 RS + 1Master)
>> and the HDFS is CDH3b3
>>
>> During the first hours of running, I puts many (tens of millions entites,
>> each about 200 bytes), it worked well.
>>
>> But then, the client cannot put more data.
>>
>> I checked all log files of hbase, no abnormal is found, I will continue to
>> check this issue.
>>
>> It seems related to ZooKeeper......
>>
>
>