Re: mapreduce job failure
400 regions a day is way too much, also in 0.20.6 there's a high risk of collision when you get near the 10 thousands of regions. But that's most probably not your current issue. That HDFS message 99% of the time means that the region server went into GC and when it came back the master already moved the regions away. Should be pretty obvious in the logs. As to why the tasks get killed, it's probably related. And since you are running such an old release you have data loss and if that happens on the .META. table then you lose metadata about the regions. To help with GC issues, I suggest you read the multi-part blog post from Todd: http://www.cloudera.com/blog/2011/02/avoiding-full-gcs-in-hbase-with-memstore-local-allocation-buffers-part-1/ J-D On Mon, May 16, 2011 at 2:08 PM, Venkatesh vramanatha...@aol.com wrote: Thanks J-D Using hbase-0.20.6, 49 node cluster The map reduce job involve a full table scan...(region size 4 gig) The job runs great for 1 week.. Starts failing after 1 week of data accumulation (about 3000 regions) About 400 regions get created per day... Can you suggest any tunables at the HBase level. or HDFS level.? Also, I've one more issue..when region servers die..Errors below: (any suggestion here is helpfull as well) org.apache.hadoop.ipc.RemoteException: org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException: No lease on /hbase_data_one_110425/.../compaction.dir/249610074/4534752250560182124 File does not exist. Holder DFSClient_-398073404 does not have any open files. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1332) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1323) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1251) at org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:422) at sun.reflect.GeneratedMethodAccessor24.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953) -Original Message- From: Jean-Daniel Cryans jdcry...@apache.org To: user@hbase.apache.org Sent: Fri, May 13, 2011 12:39 am Subject: Re: mapreduce job failure All that means is that the task stayed in map() for 10 minutes, blocked on something. If you were scanning an hbase table, and didn't get a new row after 1 minute, then the scanner would expire. That's orthogonal tho. You need to figure what you're blocking on, add logging and try to jstack your Child processes for example. J-D On Thu, May 12, 2011 at 7:21 PM, Venkatesh vramanatha...@aol.com wrote: Hi Using hbase-0.20.6 mapreduce job started failing in the map phase (using hbase table as input for mapper)..(ran fine for a week or so starting with empty tables).. task tracker log: Task attempt_201105121141_0002_m_000452_0 failed to report status for 600 seconds. Killing Region server log: 2011-05-12 18:27:39,919 INFO org.apache.hadoop.hbase.regionserver.HRegionServer: Scanner -7857209327501974146 lease expired 2011-05-12 18:28:29,716 ERROR org.apache.hadoop.hbase.regionserver.HRegionServer:org.apache.hadoop.hbase.UnknownScannerException: Name: -7857209327501974146 at org.apache.hadoop.hbase.regionserver.HRegionServer.next(HRegionServer.java:1880) at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:657) at org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:915) 2011-05-12 18:28:29,897 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server handler 3 on 60020, call next(-78572093275019 74146, 1) from .:35202: error: org.apache.hadoop.hbase.UnknownScannerException: Name: -7857209327501974146 org.apache.hadoop.hbase.UnknownScannerException: Name: -7857209327501974146 at org.apache.hadoop.hbase.regionserver.HRegionServer.next(HRegionServer.java:1880) at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597
Re: mapreduce job failure
thanks J-D as always -Original Message- From: Jean-Daniel Cryans jdcry...@apache.org To: user@hbase.apache.org Sent: Tue, May 17, 2011 8:04 pm Subject: Re: mapreduce job failure 400 regions a day is way too much, also in 0.20.6 there's a high risk of collision when you get near the 10 thousands of regions. But that's most probably not your current issue. That HDFS message 99% of the time means that the region server went into GC and when it came back the master already moved the regions away. Should be pretty obvious in the logs. As to why the tasks get killed, it's probably related. And since you are running such an old release you have data loss and if that happens on the .META. table then you lose metadata about the regions. To help with GC issues, I suggest you read the multi-part blog post from Todd: http://www.cloudera.com/blog/2011/02/avoiding-full-gcs-in-hbase-with-memstore-local-allocation-buffers-part-1/ J-D On Mon, May 16, 2011 at 2:08 PM, Venkatesh vramanatha...@aol.com wrote: Thanks J-D Using hbase-0.20.6, 49 node cluster The map reduce job involve a full table scan...(region size 4 gig) The job runs great for 1 week.. Starts failing after 1 week of data accumulation (about 3000 regions) About 400 regions get created per day... Can you suggest any tunables at the HBase level. or HDFS level.? Also, I've one more issue..when region servers die..Errors below: (any suggestion here is helpfull as well) org.apache.hadoop.ipc.RemoteException: org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException: No lease on /hbase_data_one_110425/.../compaction.dir/249610074/4534752250560182124 File does not exist. Holder DFSClient_-398073404 does not have any open files. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1332) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1323) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1251) at org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:422) at sun.reflect.GeneratedMethodAccessor24.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953) -Original Message- From: Jean-Daniel Cryans jdcry...@apache.org To: user@hbase.apache.org Sent: Fri, May 13, 2011 12:39 am Subject: Re: mapreduce job failure All that means is that the task stayed in map() for 10 minutes, blocked on something. If you were scanning an hbase table, and didn't get a new row after 1 minute, then the scanner would expire. That's orthogonal tho. You need to figure what you're blocking on, add logging and try to jstack your Child processes for example. J-D On Thu, May 12, 2011 at 7:21 PM, Venkatesh vramanatha...@aol.com wrote: Hi Using hbase-0.20.6 mapreduce job started failing in the map phase (using hbase table as input for mapper)..(ran fine for a week or so starting with empty tables).. task tracker log: Task attempt_201105121141_0002_m_000452_0 failed to report status for 600 seconds. Killing Region server log: 2011-05-12 18:27:39,919 INFO org.apache.hadoop.hbase.regionserver.HRegionServer: Scanner -7857209327501974146 lease expired 2011-05-12 18:28:29,716 ERROR org.apache.hadoop.hbase.regionserver.HRegionServer:org.apache.hadoop.hbase.UnknownScannerException: Name: -7857209327501974146 at org.apache.hadoop.hbase.regionserver.HRegionServer.next(HRegionServer.java:1880) at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source)at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:657) at org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:915) 2011-05-12 18:28:29,897 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server handler 3 on 60020, call next(-78572093275019 74146, 1) from .:35202: error: org.apache.hadoop.hbase.UnknownScannerException: Name: -7857209327501974146 org.apache.hadoop.hbase.UnknownScannerException: Name: -7857209327501974146
Re: mapreduce job failure
Thanks J-D Using hbase-0.20.6, 49 node cluster The map reduce job involve a full table scan...(region size 4 gig) The job runs great for 1 week.. Starts failing after 1 week of data accumulation (about 3000 regions) About 400 regions get created per day... Can you suggest any tunables at the HBase level. or HDFS level.? Also, I've one more issue..when region servers die..Errors below: (any suggestion here is helpfull as well) org.apache.hadoop.ipc.RemoteException: org.apache.hadoop.hdfs.server.namenode.LeaseExpiredException: No lease on /hbase_data_one_110425/.../compaction.dir/249610074/4534752250560182124 File does not exist. Holder DFSClient_-398073404 does not have any open files. at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1332) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkLease(FSNamesystem.java:1323) at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1251) at org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:422) at sun.reflect.GeneratedMethodAccessor24.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:508) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:959) at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:955) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:396) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:953) -Original Message- From: Jean-Daniel Cryans jdcry...@apache.org To: user@hbase.apache.org Sent: Fri, May 13, 2011 12:39 am Subject: Re: mapreduce job failure All that means is that the task stayed in map() for 10 minutes, blocked on something. If you were scanning an hbase table, and didn't get a new row after 1 minute, then the scanner would expire. That's orthogonal tho. You need to figure what you're blocking on, add logging and try to jstack your Child processes for example. J-D On Thu, May 12, 2011 at 7:21 PM, Venkatesh vramanatha...@aol.com wrote: Hi Using hbase-0.20.6 mapreduce job started failing in the map phase (using hbase table as input for mapper)..(ran fine for a week or so starting with empty tables).. task tracker log: Task attempt_201105121141_0002_m_000452_0 failed to report status for 600 seconds. Killing Region server log: 2011-05-12 18:27:39,919 INFO org.apache.hadoop.hbase.regionserver.HRegionServer: Scanner -7857209327501974146 lease expired 2011-05-12 18:28:29,716 ERROR org.apache.hadoop.hbase.regionserver.HRegionServer:org.apache.hadoop.hbase.UnknownScannerException: Name: -7857209327501974146 at org.apache.hadoop.hbase.regionserver.HRegionServer.next(HRegionServer.java:1880) at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source)at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:657) at org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:915) 2011-05-12 18:28:29,897 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server handler 3 on 60020, call next(-78572093275019 74146, 1) from .:35202: error: org.apache.hadoop.hbase.UnknownScannerException: Name: -7857209327501974146 org.apache.hadoop.hbase.UnknownScannerException: Name: -7857209327501974146 at org.apache.hadoop.hbase.regionserver.HRegionServer.next(HRegionServer.java:1880) at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:657) at org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:915) I don't see any error in datanodes Appreciate any help thanks v
mapreduce job failure
Hi Using hbase-0.20.6 mapreduce job started failing in the map phase (using hbase table as input for mapper)..(ran fine for a week or so starting with empty tables).. task tracker log: Task attempt_201105121141_0002_m_000452_0 failed to report status for 600 seconds. Killing Region server log: 2011-05-12 18:27:39,919 INFO org.apache.hadoop.hbase.regionserver.HRegionServer: Scanner -7857209327501974146 lease expired 2011-05-12 18:28:29,716 ERROR org.apache.hadoop.hbase.regionserver.HRegionServer:org.apache.hadoop.hbase.UnknownScannerException: Name: -7857209327501974146 at org.apache.hadoop.hbase.regionserver.HRegionServer.next(HRegionServer.java:1880) at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:657) at org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:915) 2011-05-12 18:28:29,897 INFO org.apache.hadoop.ipc.HBaseServer: IPC Server handler 3 on 60020, call next(-78572093275019 74146, 1) from .:35202: error: org.apache.hadoop.hbase.UnknownScannerException: Name: -7857209327501974146 org.apache.hadoop.hbase.UnknownScannerException: Name: -7857209327501974146 at org.apache.hadoop.hbase.regionserver.HRegionServer.next(HRegionServer.java:1880) at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at org.apache.hadoop.hbase.ipc.HBaseRPC$Server.call(HBaseRPC.java:657) at org.apache.hadoop.hbase.ipc.HBaseServer$Handler.run(HBaseServer.java:915) I don't see any error in datanodes Appreciate any help thanks v