[ https://issues.apache.org/jira/browse/HADOOP-2619?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12559294#action_12559294 ]
viper799 edited comment on HADOOP-2619 at 1/15/08 4:05 PM: ---------------------------------------------------------------- I tryed to kill the region server that the region was on and it fail over to a different server fine but compaction failed on that server too same error as above. but when I tryed to stop the whole cluster stop-hbase.sh and stop-all.sh I get this error on the new regionserver for this region {code} 2008-01-15 17:47:30,611 WARN org.apache.hadoop.fs.DFSClient: Problem renewing lease for DFSClient_647476091: java.net.ConnectException: Connection refused at java.net.PlainSocketImpl.socketConnect(Native Method) at java.net.PlainSocketImpl.doConnect(PlainSocketImpl.java:333) at java.net.PlainSocketImpl.connectToAddress(PlainSocketImpl.java:195) at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:182) at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:366) at java.net.Socket.connect(Socket.java:519) at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:159) at org.apache.hadoop.ipc.Client.getConnection(Client.java:575) at org.apache.hadoop.ipc.Client.call(Client.java:498) at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:198) at org.apache.hadoop.dfs.$Proxy1.renewLease(Unknown Source) at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:585) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59) at org.apache.hadoop.dfs.$Proxy1.renewLease(Unknown Source) at org.apache.hadoop.dfs.DFSClient$LeaseChecker.run(DFSClient.java:586) at java.lang.Thread.run(Thread.java:595) {code} I have to kill the region server my self with kill pid then on restart of the cluster I get this meta error reprted many times over and over on the region server serveing .META. {code} 2008-01-15 17:55:09,315 INFO org.apache.hadoop.ipc.Server: IPC Server handler 0 on 60020, call openScanner(.META.,,1, [Lorg.apache.hadoop.io.Text;@167e3a5, , 1200441309302, null) from 10.0.0.1:49221: error: java.io.IOException: HStoreScanner failed construction java.io.IOException: HStoreScanner failed construction at org.apache.hadoop.hbase.HStore$StoreFileScanner.<init>(HStore.java:2139) at org.apache.hadoop.hbase.HStore$HStoreScanner.<init>(HStore.java:2260) at org.apache.hadoop.hbase.HStore.getScanner(HStore.java:2050) at org.apache.hadoop.hbase.HRegion$HScanner.<init>(HRegion.java:1593) at org.apache.hadoop.hbase.HRegion.getScanner(HRegion.java:1167) at org.apache.hadoop.hbase.HRegionServer.openScanner(HRegionServer.java:1517) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:585) at org.apache.hadoop.hbase.ipc.HbaseRPC$Server.call(HbaseRPC.java:413) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:908) Caused by: org.apache.hadoop.ipc.RemoteException: java.io.IOException: File does not exist: /gfs_storage/hadoop-root/hbase/.META./1028785192/info/mapfiles/5207253989210860351/data at org.apache.hadoop.dfs.FSDirectory.getFileInfo(FSDirectory.java:543) at org.apache.hadoop.dfs.FSNamesystem.getFileInfo(FSNamesystem.java:1464) at org.apache.hadoop.dfs.NameNode.getFileInfo(NameNode.java:439) at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:585) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:409) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:908) at org.apache.hadoop.ipc.Client.call(Client.java:509) at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:198) at org.apache.hadoop.dfs.$Proxy1.getFileInfo(Unknown Source) at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:585) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59) at org.apache.hadoop.dfs.$Proxy1.getFileInfo(Unknown Source) at org.apache.hadoop.dfs.DFSClient.getFileInfo(DFSClient.java:442) at org.apache.hadoop.dfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:331) at org.apache.hadoop.fs.FileSystem.getLength(FileSystem.java:536) at org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1349) at org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1344) at org.apache.hadoop.io.MapFile$Reader.<init>(MapFile.java:254) at org.apache.hadoop.io.MapFile$Reader.<init>(MapFile.java:242) at org.apache.hadoop.hbase.HStoreFile$HbaseMapFile$HbaseReader.<init>(HStoreFile.java:595) at org.apache.hadoop.hbase.HStoreFile$BloomFilterMapFile$Reader.<init>(HStoreFile.java:650) at org.apache.hadoop.hbase.HStoreFile.getReader(HStoreFile.java:424) at org.apache.hadoop.hbase.HStore$StoreFileScanner.<init>(HStore.java:2114) ... 11 more {code} was (Author: viper799): I tryed to kill the region server that the region was on and it fail over to a different server fine but compaction failed on that server too same error as above. but when I tryed to stop the whole cluster stop-hbase.sh and stop-all.sh I get this error on the new regionserver for this region {code} 2008-01-15 17:47:30,611 WARN org.apache.hadoop.fs.DFSClient: Problem renewing lease for DFSClient_647476091: java.net.ConnectException: Connection refused at java.net.PlainSocketImpl.socketConnect(Native Method) at java.net.PlainSocketImpl.doConnect(PlainSocketImpl.java:333) at java.net.PlainSocketImpl.connectToAddress(PlainSocketImpl.java:195) at java.net.PlainSocketImpl.connect(PlainSocketImpl.java:182) at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:366) at java.net.Socket.connect(Socket.java:519) at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:159) at org.apache.hadoop.ipc.Client.getConnection(Client.java:575) at org.apache.hadoop.ipc.Client.call(Client.java:498) at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:198) at org.apache.hadoop.dfs.$Proxy1.renewLease(Unknown Source) at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:585) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59) at org.apache.hadoop.dfs.$Proxy1.renewLease(Unknown Source) at org.apache.hadoop.dfs.DFSClient$LeaseChecker.run(DFSClient.java:586) at java.lang.Thread.run(Thread.java:595) {code} then on restart of the cluster I get this meta error reprted many times over and over on the region server serveing .META. {code} 2008-01-15 17:55:09,315 INFO org.apache.hadoop.ipc.Server: IPC Server handler 0 on 60020, call openScanner(.META.,,1, [Lorg.apache.hadoop.io.Text;@167e3a5, , 1200441309302, null) from 10.0.0.1:49221: error: java.io.IOException: HStoreScanner failed construction java.io.IOException: HStoreScanner failed construction at org.apache.hadoop.hbase.HStore$StoreFileScanner.<init>(HStore.java:2139) at org.apache.hadoop.hbase.HStore$HStoreScanner.<init>(HStore.java:2260) at org.apache.hadoop.hbase.HStore.getScanner(HStore.java:2050) at org.apache.hadoop.hbase.HRegion$HScanner.<init>(HRegion.java:1593) at org.apache.hadoop.hbase.HRegion.getScanner(HRegion.java:1167) at org.apache.hadoop.hbase.HRegionServer.openScanner(HRegionServer.java:1517) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:585) at org.apache.hadoop.hbase.ipc.HbaseRPC$Server.call(HbaseRPC.java:413) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:908) Caused by: org.apache.hadoop.ipc.RemoteException: java.io.IOException: File does not exist: /gfs_storage/hadoop-root/hbase/.META./1028785192/info/mapfiles/5207253989210860351/data at org.apache.hadoop.dfs.FSDirectory.getFileInfo(FSDirectory.java:543) at org.apache.hadoop.dfs.FSNamesystem.getFileInfo(FSNamesystem.java:1464) at org.apache.hadoop.dfs.NameNode.getFileInfo(NameNode.java:439) at sun.reflect.GeneratedMethodAccessor5.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:585) at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:409) at org.apache.hadoop.ipc.Server$Handler.run(Server.java:908) at org.apache.hadoop.ipc.Client.call(Client.java:509) at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:198) at org.apache.hadoop.dfs.$Proxy1.getFileInfo(Unknown Source) at sun.reflect.GeneratedMethodAccessor4.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:585) at org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:82) at org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:59) at org.apache.hadoop.dfs.$Proxy1.getFileInfo(Unknown Source) at org.apache.hadoop.dfs.DFSClient.getFileInfo(DFSClient.java:442) at org.apache.hadoop.dfs.DistributedFileSystem.getFileStatus(DistributedFileSystem.java:331) at org.apache.hadoop.fs.FileSystem.getLength(FileSystem.java:536) at org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1349) at org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1344) at org.apache.hadoop.io.MapFile$Reader.<init>(MapFile.java:254) at org.apache.hadoop.io.MapFile$Reader.<init>(MapFile.java:242) at org.apache.hadoop.hbase.HStoreFile$HbaseMapFile$HbaseReader.<init>(HStoreFile.java:595) at org.apache.hadoop.hbase.HStoreFile$BloomFilterMapFile$Reader.<init>(HStoreFile.java:650) at org.apache.hadoop.hbase.HStoreFile.getReader(HStoreFile.java:424) at org.apache.hadoop.hbase.HStore$StoreFileScanner.<init>(HStore.java:2114) ... 11 more {code} > Compaction errors after a region splits > --------------------------------------- > > Key: HADOOP-2619 > URL: https://issues.apache.org/jira/browse/HADOOP-2619 > Project: Hadoop > Issue Type: Bug > Components: contrib/hbase > Environment: hadoop snv 612165 > Reporter: Billy Pearson > Fix For: 0.16.0 > > Attachments: hbase-root-regionserver-PE1750-4.log > > > I am getting compaction errors from regions after they split not all of them > have this problem but some do > I attached a log I picked out one region > webdata,com.technorati/tag/potiron:http,1200430376177 > it is loaded then splits at > 2008-01-15 14:52:56,116 > the split is finshed at > 2008-01-15 14:53:01,653 > the first compaction for the new top half region starts at > 2008-01-15 14:54:07,612 - > webdata,com.technorati/tag/potiron:http,1200430376177 > and ends successful at > 2008-01-15 14:54:30,229 > ten the next compaction starts at > 2008-01-15 14:56:16,315 > This one ends with an error at > 2008-01-15 14:56:40,246 > {code} > 2008-01-15 14:57:53,002 ERROR org.apache.hadoop.hbase.HRegionServer: > Compaction failed for region > webdata,com.technorati/tag/potiron:http,1200430376177 > org.apache.hadoop.dfs.LeaseExpiredException: > org.apache.hadoop.dfs.LeaseExpiredException: No lease on > /gfs_storage/hadoop-root/hbase/webdata/compaction.dir/1438658724/in_rank/mapfiles/8222904438849251562/data > at org.apache.hadoop.dfs.FSNamesystem.checkLease(FSNamesystem.java:1123) > at > org.apache.hadoop.dfs.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1061) > at org.apache.hadoop.dfs.NameNode.addBlock(NameNode.java:303) > at sun.reflect.GeneratedMethodAccessor12.invoke(Unknown Source) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > at java.lang.reflect.Method.invoke(Method.java:585) > at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:409) > at org.apache.hadoop.ipc.Server$Handler.run(Server.java:908) > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) > at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27) > at java.lang.reflect.Constructor.newInstance(Constructor.java:494) > at > org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteExceptionHandler.java:82) > at > org.apache.hadoop.hbase.RemoteExceptionHandler.checkIOException(RemoteExceptionHandler.java:48) > at > org.apache.hadoop.hbase.HRegionServer$Compactor.run(HRegionServer.java:418) > {code} > and all other compaction's for this region fail after this one fail with the > same error I will have to keep testing to see if it ever finishes > successfully. > maybe after a restart it will successfully finsh a compaction. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.