thanks J-D as always
 

 


 

 

-----Original Message-----
From: Jean-Daniel Cryans <[email protected]>
To: [email protected]
Sent: Tue, May 17, 2011 8:04 pm
Subject: Re: mapreduce job failure


400 regions a day is way too much, also in 0.20.6 there's a high risk

of collision when you get near the 10 thousands of regions. But that's

most probably not your current issue.



That HDFS message 99% of the time means that the region server went

into GC and when it came back the master already moved the regions

away. Should be pretty obvious in the logs.



As to why the tasks get killed, it's probably related. And since you

are running such an old release you have data loss and if that happens

on the .META. table then you lose metadata about the regions.



To help with GC issues, I suggest you read the multi-part blog post

from Todd: 
http://www.cloudera.com/blog/2011/02/avoiding-full-gcs-in-hbase-with-memstore-local-allocation-buffers-part-1/



J-D



On Mon, May 16, 2011 at 2:08 PM, Venkatesh <[email protected]> wrote:

> Thanks J-D

>

> Using hbase-0.20.6, 49 node cluster

>

> The map reduce job involve a full table scan...(region size 4 gig)

> The job runs great for 1 week..

> Starts failing after 1 week of data accumulation (about 3000 regions)

>

>  About 400 regions get created per day...

>

> Can you suggest any tunables at the HBase level. or HDFS level.?

>

> Also, I've one more issue..when region servers die..Errors below: (any 

suggestion here is helpfull as well)

>

> org.apache.hadoop.ipc.RemoteException: 
> org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException: 

No lease on 
/hbase_data_one_110425/.../compaction.dir/249610074/4534752250560182124 

File does not exist. Holder DFSClient_-398073404 does not have any open files.

>        at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1332)

>        at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1323)

>        at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1251)

>        at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:422)

>        at sun.reflect.GeneratedMethodAccessor24.invoke(Unknown Source)

>        at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)

>        at java.lang.reflect.Method.invoke(Method.java:597)

>        at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508)

>        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959)

>        at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955)

>        at java.security.AccessController.doPrivileged(Native Method)

>        at javax.security.auth.Subject.doAs(Subject.java:396)

>        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953)

>

>

>

>

>

>

>

>

>

> -----Original Message-----

> From: Jean-Daniel Cryans <[email protected]>

> To: [email protected]

> Sent: Fri, May 13, 2011 12:39 am

> Subject: Re: mapreduce job failure

>

>

> All that means is that the task stayed in map() for 10 minutes,

>

> blocked on "something".

>

>

>

> If you were scanning an hbase table, and didn't get a new row after 1

>

> minute, then the scanner would expire. That's orthogonal tho.

>

>

>

> You need to figure what you're blocking on, add logging and try to

>

> jstack your Child processes for example.

>

>

>

> J-D

>

>

>

> On Thu, May 12, 2011 at 7:21 PM, Venkatesh <[email protected]> wrote:

>

>>

>

>>  Hi

>

>> Using hbase-0.20.6

>

>>

>

>> mapreduce job started failing in the map phase (using hbase table as input 

for

>

> mapper)..(ran fine for a week or so starting with empty tables)..

>

>>

>

>> task tracker log:

>

>>

>

>>

>

>> Task attempt_201105121141_0002_m_000452_0 failed to report status for 600

>

> seconds. Killing

>

>>

>

>>

>

>> Region server log:

>

>>

>

>> 2011-05-12 18:27:39,919 INFO 
>> org.apache.hadoop.hbase.regionserver.HRegionServer:

>

> Scanner -7857209327501974146 lease expired

>

>>

>

>> 2011-05-12 18:28:29,716 ERROR 
>> org.apache.hadoop.hbase.regionserver.HRegionServer:org.apache.hadoop.hbase.UnknownScannerException:

>

> Name: -7857209327501974146

>

>>        at 
>> org.apache.hadoop.hbase.regionserver.HRegionServer.next(HRegionServer.java:1880)

>

>       at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source)        at

>

> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)

>

>>        at java.lang.reflect.Method.invoke(Method.java:597)

>

>>        at org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:657)

>

>>        at 
>> org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:915)

>

>>

>

>>

>

>> 2011-05-12 18:28:29,897 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server

>

> handler 3 on 60020, call next(-78572093275019

>

>> 74146, 1) from .....:35202: error: 
>> org.apache.hadoop.hbase.UnknownScannerException:

>

> Name: -7857209327501974146

>

>> org.apache.hadoop.hbase.UnknownScannerException: Name: -7857209327501974146

>

>>        at 
>> org.apache.hadoop.hbase.regionserver.HRegionServer.next(HRegionServer.java:1880)

>

>>        at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source)

>

>>        at 
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)

>

>       at java.lang.reflect.Method.invoke(Method.java:597)

>

>>        at org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:657)

>

>>        at 
>> org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:915)

>

>>

>

>> I don't see any error in datanodes

>

>>

>

>> Appreciate any help

>

>> thanks

>

>> v

>

>>

>

>>

>

>>

>

>

>

>


 
=

Reply via email to