I can't be definitive, but I have had a very similar issue in the past.  The 
root cause was the my NTP server had died and a couple of nodes in the cluster 
got wildly out of sync.  Check your HDFS health And if there are 
under-replicated blocks...this "could" be your issue (though root cause could 
be bad disks or any number of other issues that present with the same 
"symptoms", but again, I would take this advice only as far as needed to either 
rule it out or dig further...don't go down a rabbit hole.  Your errors could 
have been caused by an entirely different problem...I have no other context 
from the error you provided to know where else to look...

Aaron

Sent from my iPhone

> On Jul 6, 2017, at 16:55, Ted Yu <yuzhih...@gmail.com> wrote:
> 
> HBASE-16960 mentioned the following :
> 
> Caused by: java.net.SocketTimeoutException: 20000 millis timeout while 
> waiting for channel to be ready for read
> 
> Do you see similar line in region server log ?
> 
> Cheers
> 
>> On Thu, Jul 6, 2017 at 1:48 PM, anil gupta <anilgupt...@gmail.com> wrote:
>> Hi All,
>> 
>> We are running HBase/Phoenix on EMR5.2(HBase1.2.3 and Phoenix4.7) and we 
>> running into following exception when we are trying to load data into one of 
>> our Phoenix table:
>> 2017-07-06 19:57:57,507 INFO [hconnection-0x60e5272-shared--pool2-t249] 
>> org.apache.hadoop.hbase.client.AsyncProcess: #1, table=DE.CONFIG_DATA, 
>> attempt=30/35 failed=38ops, last exception: 
>> org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: 
>> org.apache.hadoop.hbase.regionserver.wal.DamagedWALException: Append 
>> sequenceId=8689, requesting roll of WAL
>>      at 
>> org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.append(FSHLog.java:1921)
>>      at 
>> org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1773)
>>      at 
>> org.apache.hadoop.hbase.regionserver.wal.FSHLog$RingBufferEventHandler.onEvent(FSHLog.java:1695)
>>      at 
>> com.lmax.disruptor.BatchEventProcessor.run(BatchEventProcessor.java:128)
>>      at 
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
>>      at 
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
>>      at java.lang.Thread.run(Thread.java:745)
>> 
>> We are OK with wiping out this table and rebuilding the dataset. We tried to 
>> drop the table and recreate the table but it didnt fix it. 
>> Can anyone please let us know how can we get rid of above problem? Are we 
>> running into https://issues.apache.org/jira/browse/HBASE-16960?
>> 
>> -- 
>> Thanks & Regards,
>> Anil Gupta
> 

Reply via email to