[ 
https://issues.apache.org/jira/browse/HADOOP-2619?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12559294#action_12559294
 ] 

viper799 edited comment on HADOOP-2619 at 1/15/08 4:05 PM:
----------------------------------------------------------------

I tryed to kill the region server that the region was on and it fail over to a 
different server fine but compaction failed on that server too same error as 
above.

but when I tryed to stop the whole cluster stop-hbase.sh and stop-all.sh I get 
this error on the new regionserver for this region

{code}
2008-01-15 17:47:30,611 WARN org.apache.hadoop.fs.DFSClient: Problem renewing 
lease for DFSClient_647476091: java.net.ConnectException: Connection refused
        at java.net.PlainSocketImpl.socketConnect(Native Method)
        at java.net.PlainSocketImpl.doConnect(PlainSocketImpl.java:333)
        at java.net.PlainSocketImpl.connectToAddress(PlainSocketImpl.java:195)
        at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:182)
        at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:366)
        at java.net.Socket.connect(Socket.java:519)
        at 
org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:159)
        at org.apache.hadoop.ipc.Client.getConnection(Client.java:575)
        at org.apache.hadoop.ipc.Client.call(Client.java:498)
        at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:198)
        at org.apache.hadoop.dfs.$Proxy1.renewLease(Unknown Source)
        at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:585)
        at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
        at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
        at org.apache.hadoop.dfs.$Proxy1.renewLease(Unknown Source)
        at org.apache.hadoop.dfs.DFSClient$LeaseChecker.run(DFSClient.java:586)
        at java.lang.Thread.run(Thread.java:595)

{code}
I have to kill the region server my self with kill pid
then on restart of the cluster I get this meta error reprted many times over 
and over on the region server serveing .META.

{code}
2008-01-15 17:55:09,315 INFO org.apache.hadoop.ipc.Server: IPC Server handler 0 
on 60020, call openScanner(.META.,,1, [Lorg.apache.hadoop.io.Text;@167e3a5, , 
1200441309302, null) from 10.0.0.1:49221: error: java.io.IOException: 
HStoreScanner failed construction
java.io.IOException: HStoreScanner failed construction
        at 
org.apache.hadoop.hbase.HStore$StoreFileScanner.<init>(HStore.java:2139)
        at org.apache.hadoop.hbase.HStore$HStoreScanner.<init>(HStore.java:2260)
        at org.apache.hadoop.hbase.HStore.getScanner(HStore.java:2050)
        at org.apache.hadoop.hbase.HRegion$HScanner.<init>(HRegion.java:1593)
        at org.apache.hadoop.hbase.HRegion.getScanner(HRegion.java:1167)
        at 
org.apache.hadoop.hbase.HRegionServer.openScanner(HRegionServer.java:1517)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:585)
        at org.apache.hadoop.hbase.ipc.HbaseRPC$Server.call(HbaseRPC.java:413)
        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:908)
Caused by: org.apache.hadoop.ipc.RemoteException: java.io.IOException: File 
does not exist: 
/gfs_storage/hadoop-root/hbase/.META./1028785192/info/mapfiles/5207253989210860351/data
        at org.apache.hadoop.dfs.FSDirectory.getFileInfo(FSDirectory.java:543)
        at 
org.apache.hadoop.dfs.FSNamesystem.getFileInfo(FSNamesystem.java:1464)
        at org.apache.hadoop.dfs.NameNode.getFileInfo(NameNode.java:439)
        at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:585)
        at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:409)
        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:908)

        at org.apache.hadoop.ipc.Client.call(Client.java:509)
        at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:198)
        at org.apache.hadoop.dfs.$Proxy1.getFileInfo(Unknown Source)
        at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:585)
        at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
        at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
        at org.apache.hadoop.dfs.$Proxy1.getFileInfo(Unknown Source)
        at org.apache.hadoop.dfs.DFSClient.getFileInfo(DFSClient.java:442)
        at 
org.apache.hadoop.dfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:331)
        at org.apache.hadoop.fs.FileSystem.getLength(FileSystem.java:536)
        at 
org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1349)
        at 
org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1344)
        at org.apache.hadoop.io.MapFile$Reader.<init>(MapFile.java:254)
        at org.apache.hadoop.io.MapFile$Reader.<init>(MapFile.java:242)
        at 
org.apache.hadoop.hbase.HStoreFile$HbaseMapFile$HbaseReader.<init>(HStoreFile.java:595)
        at 
org.apache.hadoop.hbase.HStoreFile$BloomFilterMapFile$Reader.<init>(HStoreFile.java:650)
        at org.apache.hadoop.hbase.HStoreFile.getReader(HStoreFile.java:424)
        at 
org.apache.hadoop.hbase.HStore$StoreFileScanner.<init>(HStore.java:2114)
        ... 11 more

{code}

      was (Author: viper799):
    I tryed to kill the region server that the region was on and it fail over 
to a different server fine but compaction failed on that server too same error 
as above.

but when I tryed to stop the whole cluster stop-hbase.sh and stop-all.sh I get 
this error on the new regionserver for this region

{code}
2008-01-15 17:47:30,611 WARN org.apache.hadoop.fs.DFSClient: Problem renewing 
lease for DFSClient_647476091: java.net.ConnectException: Connection refused
        at java.net.PlainSocketImpl.socketConnect(Native Method)
        at java.net.PlainSocketImpl.doConnect(PlainSocketImpl.java:333)
        at java.net.PlainSocketImpl.connectToAddress(PlainSocketImpl.java:195)
        at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:182)
        at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:366)
        at java.net.Socket.connect(Socket.java:519)
        at 
org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:159)
        at org.apache.hadoop.ipc.Client.getConnection(Client.java:575)
        at org.apache.hadoop.ipc.Client.call(Client.java:498)
        at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:198)
        at org.apache.hadoop.dfs.$Proxy1.renewLease(Unknown Source)
        at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:585)
        at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
        at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
        at org.apache.hadoop.dfs.$Proxy1.renewLease(Unknown Source)
        at org.apache.hadoop.dfs.DFSClient$LeaseChecker.run(DFSClient.java:586)
        at java.lang.Thread.run(Thread.java:595)

{code}

then on restart of the cluster I get this meta error reprted many times over 
and over on the region server serveing .META.

{code}
2008-01-15 17:55:09,315 INFO org.apache.hadoop.ipc.Server: IPC Server handler 0 
on 60020, call openScanner(.META.,,1, [Lorg.apache.hadoop.io.Text;@167e3a5, , 
1200441309302, null) from 10.0.0.1:49221: error: java.io.IOException: 
HStoreScanner failed construction
java.io.IOException: HStoreScanner failed construction
        at 
org.apache.hadoop.hbase.HStore$StoreFileScanner.<init>(HStore.java:2139)
        at org.apache.hadoop.hbase.HStore$HStoreScanner.<init>(HStore.java:2260)
        at org.apache.hadoop.hbase.HStore.getScanner(HStore.java:2050)
        at org.apache.hadoop.hbase.HRegion$HScanner.<init>(HRegion.java:1593)
        at org.apache.hadoop.hbase.HRegion.getScanner(HRegion.java:1167)
        at 
org.apache.hadoop.hbase.HRegionServer.openScanner(HRegionServer.java:1517)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:585)
        at org.apache.hadoop.hbase.ipc.HbaseRPC$Server.call(HbaseRPC.java:413)
        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:908)
Caused by: org.apache.hadoop.ipc.RemoteException: java.io.IOException: File 
does not exist: 
/gfs_storage/hadoop-root/hbase/.META./1028785192/info/mapfiles/5207253989210860351/data
        at org.apache.hadoop.dfs.FSDirectory.getFileInfo(FSDirectory.java:543)
        at 
org.apache.hadoop.dfs.FSNamesystem.getFileInfo(FSNamesystem.java:1464)
        at org.apache.hadoop.dfs.NameNode.getFileInfo(NameNode.java:439)
        at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:585)
        at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:409)
        at org.apache.hadoop.ipc.Server$Handler.run(Server.java:908)

        at org.apache.hadoop.ipc.Client.call(Client.java:509)
        at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:198)
        at org.apache.hadoop.dfs.$Proxy1.getFileInfo(Unknown Source)
        at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source)
        at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:585)
        at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82)
        at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59)
        at org.apache.hadoop.dfs.$Proxy1.getFileInfo(Unknown Source)
        at org.apache.hadoop.dfs.DFSClient.getFileInfo(DFSClient.java:442)
        at 
org.apache.hadoop.dfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:331)
        at org.apache.hadoop.fs.FileSystem.getLength(FileSystem.java:536)
        at 
org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1349)
        at 
org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1344)
        at org.apache.hadoop.io.MapFile$Reader.<init>(MapFile.java:254)
        at org.apache.hadoop.io.MapFile$Reader.<init>(MapFile.java:242)
        at 
org.apache.hadoop.hbase.HStoreFile$HbaseMapFile$HbaseReader.<init>(HStoreFile.java:595)
        at 
org.apache.hadoop.hbase.HStoreFile$BloomFilterMapFile$Reader.<init>(HStoreFile.java:650)
        at org.apache.hadoop.hbase.HStoreFile.getReader(HStoreFile.java:424)
        at 
org.apache.hadoop.hbase.HStore$StoreFileScanner.<init>(HStore.java:2114)
        ... 11 more

{code}
  
> Compaction errors after a region splits
> ---------------------------------------
>
>                 Key: HADOOP-2619
>                 URL: https://issues.apache.org/jira/browse/HADOOP-2619
>             Project: Hadoop
>          Issue Type: Bug
>          Components: contrib/hbase
>         Environment: hadoop snv 612165
>            Reporter: Billy Pearson
>             Fix For: 0.16.0
>
>         Attachments: hbase-root-regionserver-PE1750-4.log
>
>
> I am getting compaction errors from regions after they split not all of them 
> have this problem but some do
> I attached a log I picked out one region 
> webdata,com.technorati/tag/potiron:http,1200430376177
> it is loaded then splits at 
> 2008-01-15 14:52:56,116
> the split is finshed at
> 2008-01-15 14:53:01,653
> the first compaction for the new top half region starts at
> 2008-01-15 14:54:07,612 - 
> webdata,com.technorati/tag/potiron:http,1200430376177
> and ends successful at
> 2008-01-15 14:54:30,229
> ten the next compaction starts at
> 2008-01-15 14:56:16,315
> This one ends with an error at 
> 2008-01-15 14:56:40,246
> {code}
> 2008-01-15 14:57:53,002 ERROR org.apache.hadoop.hbase.HRegionServer: 
> Compaction failed for region 
> webdata,com.technorati/tag/potiron:http,1200430376177
> org.apache.hadoop.dfs.LeaseExpiredException: 
> org.apache.hadoop.dfs.LeaseExpiredException: No lease on 
> /gfs_storage/hadoop-root/hbase/webdata/compaction.dir/1438658724/in_rank/mapfiles/8222904438849251562/data
>       at org.apache.hadoop.dfs.FSNamesystem.checkLease(FSNamesystem.java:1123)
>       at 
> org.apache.hadoop.dfs.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1061)
>       at org.apache.hadoop.dfs.NameNode.addBlock(NameNode.java:303)
>       at sun.reflect.GeneratedMethodAccessor12.invoke(Unknown Source)
>       at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>       at java.lang.reflect.Method.invoke(Method.java:585)
>       at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:409)
>       at org.apache.hadoop.ipc.Server$Handler.run(Server.java:908)
>       at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
>       at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
>       at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
>       at java.lang.reflect.Constructor.newInstance(Constructor.java:494)
>       at 
> org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteExceptionHandler.java:82)
>       at 
> org.apache.hadoop.hbase.RemoteExceptionHandler.checkIOException(RemoteExceptionHandler.java:48)
>       at 
> org.apache.hadoop.hbase.HRegionServer$Compactor.run(HRegionServer.java:418)
> {code}
> and all other compaction's for this region fail after this one fail with the 
> same error I will have to keep testing to see if it ever finishes 
> successfully. 
> maybe after a restart it will successfully finsh a compaction.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to