Yossi,

I wrote a comment in 1094, let's solve this as fast as possible so you can
continue using HBase.

Also, a personal question, why using 5 Datanodes with 2 HRegionServers? In a
"normal" setup, those numbers would be the same.

Thx,

J-D

On Thu, Dec 25, 2008 at 2:55 AM, Yossi Ittach <[email protected]> wrote:

> Hi All
>
> After inserting a a couple of million files , I get these Errors and I
> can't
> insert anymore files. The Master seems to be OK(attached) and so does sthe
> RegionServer (also attached , but not interesting)
>
>  Has anybody encountered something like this?
>
> *(Console)*
> org.apache.hadoop.hbase.NotServingRegionException:
> org.apache.hadoop.hbase.NotServingRegionException: Region
> test2,a521DfAPKkUbWqIOHc8pAQ==,1230151003797 closed
>        at
>
> org.apache.hadoop.hbase.regionserver.HRegion.obtainRowLock(HRegion.java:1810)
>        at
> org.apache.hadoop.hbase.regionserver.HRegion.getLock(HRegion.java:1875)
>        at
> org.apache.hadoop.hbase.regionserver.HRegion.batchUpdate(HRegion.java:1406)
>        at
> org.apache.hadoop.hbase.regionserver.HRegion.batchUpdate(HRegion.java:1380)
>        at
>
> org.apache.hadoop.hbase.regionserver.HRegionServer.batchUpdate(HRegionServer.java:1114)
>        at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source)
>        at
>
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>        at java.lang.reflect.Method.invoke(Method.java:585)
>        at
> org.apache.hadoop.hbase.ipc.HbaseRPC$Server.call(HbaseRPC.java:554)
>        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:888)
>
> org.apache.hadoop.hbase.NotServingRegionException:
> org.apache.hadoop.hbase.NotServingRegionException: Region
> test2,a521DfAPKkUbWqIOHc8pAQ==,1230151003797 closed
>        at
>
> org.apache.hadoop.hbase.regionserver.HRegion.obtainRowLock(HRegion.java:1810)
>        at
> org.apache.hadoop.hbase.regionserver.HRegion.getLock(HRegion.java:1875)
>        at
> org.apache.hadoop.hbase.regionserver.HRegion.batchUpdate(HRegion.java:1406)
>        at
> org.apache.hadoop.hbase.regionserver.HRegion.batchUpdate(HRegion.java:1380)
>        at
>
> org.apache.hadoop.hbase.regionserver.HRegionServer.batchUpdate(HRegionServer.java:1114)
>        at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source)
>        at
>
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>        at java.lang.reflect.Method.invoke(Method.java:585)
>        at
> org.apache.hadoop.hbase.ipc.HbaseRPC$Server.call(HbaseRPC.java:554)
>        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:888)
>
>
>        at
>
> org.apache.hadoop.hbase.client.HConnectionManager$TableServers.getRegionServerWithRetries(HConnectionManager.java:863)
>        at org.apache.hadoop.hbase.client.HTable.commit(HTable.java:964)
>        at org.apache.hadoop.hbase.client.HTable.commit(HTable.java:950)
>        at
>
> com.outbrain.globals.io.filesystem.HBaseFeedEntries.saveToMechanisem(HBaseFeedEntries.java:137)
>        at
>
> com.outbrain.globals.io.filesystem.HBaseFeedEntries.saveTo(HBaseFeedEntries.java:108)
>        at
>
> com.outbrain.BatchFeedInserter.BatchFeedInserter$DocFeeder.call(BatchFeedInserter.java:98)
>        at
>
> com.outbrain.BatchFeedInserter.BatchFeedInserter$DocFeeder.call(BatchFeedInserter.java:1)
>        at
> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:269)
>        at java.util.concurrent.FutureTask.run(FutureTask.java:123)
>        at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:417)
>        at
> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:269)
>        at java.util.concurrent.FutureTask.run(FutureTask.java:123)
>        at
>
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:65)
>        at
>
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:168)
>        at
>
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:650)
>        at
>
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:675)
>        at java.lang.Thread.run(Thread.java:595)
>
> *(Master*)
> 2008-12-24 17:28:13,756 DEBUG org.apache.hadoop.hbase.master.BaseScanner:
> RegionManager.metaScanner REGION => {NAME =>
> 'test2,yyPOB0kIPFX7lx9pkw7Hkw==,1230147405336', STARTKEY =>
> 'yyPOB0kIPFX7lx9pkw7Hkw==', ENDKEY => '', ENCODED => 1385803844, TABLE =>
> {{NAME => 'test2', IS_ROOT => 'false', IS_META => 'false', COMPRESSION =>
> 'RECORD', FAMILIES => [{NAME => 'obde_content', BLOOMFILTER => 'false',
> IN_MEMORY => 'false', VERSIONS => '3', BLOCKCACHE => 'false', LENGTH =>
> '2147483647', TTL => '-1', COMPRESSION => 'NONE'}]}}}, SERVER => '
> 192.168.252.213:60020', STARTCODE => 1230135899203
> 2008-12-24 17:28:13,757 INFO org.apache.hadoop.hbase.master.BaseScanner:
> RegionManager.metaScanner scan of meta region {regionname: .META.,,1,
> startKey: <>, server: 192.168.252.213:60020} complete
> 2008-12-24 17:28:13,757 INFO org.apache.hadoop.hbase.master.BaseScanner:
> all
> meta regions scanned
> 2008-12-24 17:28:14,430 DEBUG org.apache.hadoop.hbase.master.ServerManager:
> Total Load: 66, Num Servers: 1, Avg Load: 66.0
> 2008-12-24 17:28:29,448 DEBUG org.apache.hadoop.hbase.master.ServerManager:
> Total Load: 66, Num Servers: 1, Avg Load: 66.0
> 2008-12-24 17:28:38,355 DEBUG
> org.apache.hadoop.hbase.client.HConnectionManager$TableServers: Cache hit
> in
> table locations for row <> and tableName .META.: location server
> 192.168.252.213:60020, location region name .META.,,1
> 2008-12-24 17:28:44,465 DEBUG org.apache.hadoop.hbase.master.ServerManager:
> Total Load: 66, Num Servers: 1, Avg Load: 66.0
> 2008-12-24 17:28:59,483 DEBUG org.apache.hadoop.hbase.master.ServerManager:
> Total Load: 66, Num Servers: 1, Avg Load: 66.0
> 2008-12-24 17:29:07,706 INFO org.apache.hadoop.hbase.master.BaseScanner:
> RegionManager.rootScanner scanning meta region {regionname: -ROOT-,,0,
> startKey: <>, server: 192.168.252.213:60020}
> 2008-12-24 17:29:07,722 DEBUG org.apache.hadoop.hbase.master.BaseScanner:
> RegionManager.rootScanner REGION => {NAME => '.META.,,1', STARTKEY => '',
> ENDKEY => '', ENCODED => 1028785192, TABLE => {{NAME => '.META.', IS_ROOT
> =>
> 'false', IS_META => 'true', FAMILIES => [{NAME => 'historian', BLOOMFILTER
> => 'false', IN_MEMORY => 'false', VERSIONS => '2147483647', BLOCKCACHE =>
> 'false', LENGTH => '2147483647', TTL => '-1', COMPRESSION => 'NONE'}, {NAME
> => 'info', BLOOMFILTER => 'false', IN_MEMORY => 'false', VERSIONS => '1',
> BLOCKCACHE => 'false', LENGTH => '2147483647', TTL => '-1', COMPRESSION =>
> 'NONE'}]}}}, SERVER => '192.168.252.213:60020', STARTCODE => 1230135899203
> 2008-12-24 17:29:07,723 INFO org.apache.hadoop.hbase.master.BaseScanner:
> RegionManager.rootScanner scan of meta region {regionname: -ROOT-,,0,
> startKey: <>, server: 192.168.252.213:60020} complete
> 2008-12-24 17:29:09,215 DEBUG
> org.apache.hadoop.hbase.client.HConnectionManager$TableServers: Cache hit
> in
> table locations for row <> and tableName .META.: location server
> 192.168.252.213:60020, location region name .META.,,1
> 2008-12-24 17:29:13,658 INFO org.apache.hadoop.hbase.master.BaseScanner:
> RegionManager.metaScanner scanning meta region {regionname: .META.,,1,
> startKey: <>, server: 192.168.252.213:60020}
> 2008-12-24 17:29:13,685 DEBUG org.apache.hadoop.hbase.master.BaseScanner:
> RegionManager.metaScanner REGION => {NAME => 'test2,,1230150960296',
> STARTKEY => '', ENDKEY => '+zNfXoK2KxY3/ZVR5Ko4Tw==', ENCODED => 934049166,
> TABLE => {{NAME => 'test2', IS_ROOT => 'false', IS_META => 'false',
> COMPRESSION => 'RECORD', FAMILIES => [{NAME => 'obde_content', BLOOMFILTER
> => 'false', IN_MEMORY => 'false', VERSIONS => '3', BLOCKCACHE => 'false',
> LENGTH => '2147483647', TTL => '-1', COMPRESSION => 'NONE'}]}}}, SERVER =>
> '
> 192.168.252.213:60020', STARTCODE => 1230135899203
> 2008-12-24 17:29:13,686 DEBUG org.apache.hadoop.hbase.master.BaseScanner:
> RegionManager.metaScanner REGION => {NAME =>
> 'test2,+zNfXoK2KxY3/ZVR5Ko4Tw==,1230150960296', STARTKEY =>
> '+zNfXoK2KxY3/ZVR5Ko4Tw==', ENDKEY => '/z7MIyWkSwKeEUpzP1nr/w==', ENCODED
> =>
> 966894266, TABLE => {{NAME => 'test2', IS_ROOT => 'false', IS_META =>
> 'false', COMPRESSION => 'RECORD', FAMILIES => [{NAME => 'obde_content',
> BLOOMFILTER => 'false', IN_MEMORY => 'false', VERSIONS => '3', BLOCKCACHE
> =>
> 'false', LENGTH => '2147483647', TTL => '-1', COMPRESSION => 'NONE'}]}}},
> SERVER => '192.168.252.213:60020', STARTCODE => 1230135899203
> 2008-12-24 17:29:13,687 DEBUG org.apache.hadoop.hbase.master.BaseScanner:
> RegionManager.metaScanner REGION => {NAME =>
> 'test2,/z7MIyWkSwKeEUpzP1nr/w==,1230150955382', STARTKEY =>
> '/z7MIyWkSwKeEUpzP1nr/w==', ENDKEY => '0zI93xH77rV7n0ELh7/abw==', ENCODED
> =>
> 1597194225, TABLE => {{NAME => 'test2', IS_ROOT => 'false', IS_META =>
> 'false', COMPRESSION => 'RECORD', FAMILIES => [{NAME => 'obde_content',
> BLOOMFILTER => 'false', IN_MEMORY => 'false', VERSIONS => '3', BLOCKCACHE
> =>
> 'false', LENGTH => '2147483647', TTL => '-1', COMPRESSION => 'NONE'}]}}},
> SERVER => '192.168.252.213:60020', STARTCODE => 1230135899203
> 2008-12-24 17:29:13,688 DEBUG org.apache.hadoop.hbase.master.BaseScanner:
> RegionManager.metaScanner REGION => {NAME =>
> 'test2,0zI93xH77rV7n0ELh7/abw==,1230150955382', STARTKEY =>
> '0zI93xH77rV7n0ELh7/abw==', ENDKEY => '1yfvsSuQanvqzPLiozVHcw==', ENCODED
> =>
> 1323416154, TABLE => {{NAME => 'test2', IS_ROOT => 'false', IS_META =>
> 'false', COMPRESSION => 'RECORD', FAMILIES => [{NAME => 'obde_content',
> BLOOMFILTER => 'false', IN_MEMORY => 'false', VERSIONS => '3', BLOCKCACHE
>
> *RegionServer*
> ck:9558528 lastPacketInBlock:false
> 2008-12-24 17:28:45,255 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> writeChunk allocating new packet 148
> 2008-12-24 17:28:45,257 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> received ack for seqno 147
> 2008-12-24 17:28:46,563 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> writeChunk packet full seqno 148
> 2008-12-24 17:28:46,563 DEBUG org.apache.hadoop.dfs.DFSClient: DataStreamer
> block blk_5412428864980798884_22695 wrote packet seqno:148 size:65557
> offsetInBlock:9623552 lastPacketInBlock:false
> 2008-12-24 17:28:46,563 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> writeChunk allocating new packet 149
> 2008-12-24 17:28:46,566 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> received ack for seqno 148
> 2008-12-24 17:28:47,439 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> writeChunk packet full seqno 149
> 2008-12-24 17:28:47,439 DEBUG org.apache.hadoop.dfs.DFSClient: DataStreamer
> block blk_5412428864980798884_22695 wrote packet seqno:149 size:65557
> offsetInBlock:9688576 lastPacketInBlock:false
> 2008-12-24 17:28:47,439 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> writeChunk allocating new packet 150
> 2008-12-24 17:28:47,442 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> received ack for seqno 149
> 2008-12-24 17:28:48,831 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> writeChunk packet full seqno 150
> 2008-12-24 17:28:48,832 DEBUG org.apache.hadoop.dfs.DFSClient: DataStreamer
> block blk_5412428864980798884_22695 wrote packet seqno:150 size:65557
> offsetInBlock:9753600 lastPacketInBlock:false
> 2008-12-24 17:28:48,832 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> writeChunk allocating new packet 151
> 2008-12-24 17:28:48,835 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> received ack for seqno 150
> 2008-12-24 17:28:50,109 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> writeChunk packet full seqno 151
> 2008-12-24 17:28:50,109 DEBUG org.apache.hadoop.dfs.DFSClient: DataStreamer
> block blk_5412428864980798884_22695 wrote packet seqno:151 size:65557
> offsetInBlock:9818624 lastPacketInBlock:false
> 2008-12-24 17:28:50,109 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> writeChunk allocating new packet 152
> 2008-12-24 17:28:50,112 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> received ack for seqno 151
> 2008-12-24 17:28:50,942 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> writeChunk packet full seqno 152
> 2008-12-24 17:28:50,942 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> writeChunk allocating new packet 153
> 2008-12-24 17:28:50,942 DEBUG org.apache.hadoop.dfs.DFSClient: DataStreamer
> block blk_5412428864980798884_22695 wrote packet seqno:152 size:65557
> offsetInBlock:9883648 lastPacketInBlock:false
> 2008-12-24 17:28:50,944 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> received ack for seqno 152
> 2008-12-24 17:28:51,858 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> writeChunk packet full seqno 153
> 2008-12-24 17:28:51,858 DEBUG org.apache.hadoop.dfs.DFSClient: DataStreamer
> block blk_5412428864980798884_22695 wrote packet seqno:153 size:65557
> offsetInBlock:9948672 lastPacketInBlock:false
> 2008-12-24 17:28:51,858 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> writeChunk allocating new packet 154
> 2008-12-24 17:28:51,861 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> received ack for seqno 153
> 2008-12-24 17:28:54,004 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> writeChunk packet full seqno 154
> 2008-12-24 17:28:54,004 DEBUG org.apache.hadoop.dfs.DFSClient: DataStreamer
> block blk_5412428864980798884_22695 wrote packet seqno:154 size:65557
> offsetInBlock:10013696 lastPacketInBlock:false
> 2008-12-24 17:28:54,004 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> writeChunk allocating new packet 155
> 2008-12-24 17:28:54,007 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> received ack for seqno 154
> 2008-12-24 17:28:54,709 DEBUG org.apache.hadoop.dfs.DFSClient: DFSClient
> writeChunk packet full seqno 155
> 2008-12-24 17:28:54,709 DEBUG org.apache.hadoop.dfs.DFSClient: DataStreamer
> block blk_5412428864980798884_22695 wrote packet seqno:155 size:65557
> offsetInBlock:10078720 lastPacketInBlock:false
>
>
>
> Vale et me ama
> Yossi
>

Reply via email to