OK, you can get the log from http://pan.baidu.com/s/1pqS6E
2015-05-16 13:26 GMT+08:00 Ted Yu <[email protected]>: > Can you check server log on 192.168.39.22 > <http://l-namenode2.dba.cn8.qunar.com/192.168.39.22:60000> ? > > That should give you some clue. > > Cheers > > On Fri, May 15, 2015 at 8:22 PM, Louis Hust <[email protected]> wrote: > > > Hi all, > > > > I use hbase0.96.0 with hadoop 2.2.0, > > and the custom said they can not write into hbase cluster, > > So i stop the HMaster and start it soon, > > > > But it seems that the HMaster not response to request, following is the > > HMaster log: > > > > {logs} > > 2015-05-15 12:13:33,136 INFO [AM.ZK.Worker-pool2-t16] > master.RegionStates: > > Transitioned {9036a3befee90eeffb9082f90a4a9afa state=OPENING, > > ts=1431663212637, server=l-hbase26.data.cn8.qunar.com > ,60020,1431462615651} > > to {9036a3befee90eeffb9082f90a4a9afa state=OPEN, ts=1431663213136, > server= > > l-hbase26.data.cn8.qunar.com,60020,1431462615651} > > 2015-05-15 12:13:33,139 INFO [AM.ZK.Worker-pool2-t4] > master.RegionStates: > > Onlined 9036a3befee90eeffb9082f90a4a9afa on l-hbase26.data.cn8.qunar.com > > ,60020,1431462615651 > > 2015-05-15 12:14:40,699 INFO [main-EventThread] > > zookeeper.RegionServerTracker: RegionServer ephemeral node deleted, > > processing expiration [l-hbase27.data.cn8.qunar.com,60020,1431663208899] > > 2015-05-15 12:15:04,899 INFO [main-EventThread] > > zookeeper.RegionServerTracker: RegionServer ephemeral node deleted, > > processing expiration [l-hbase25.data.cn8.qunar.com,60020,1431663193865] > > 2015-05-15 12:15:24,465 WARN [249240421@qtp-591022857-33] > > client.HConnectionManager$HConnectionImplementation: Checking master > > connection > > com.google.protobuf.ServiceException: java.net.SocketTimeoutException: > Call > > to l-namenode2.dba.cn8.qunar.com/192.168.39.22:60000 failed because > > java.net.SocketTimeoutException: 60000 millis timeout while waiting for > > channel to be ready for read. ch : > > java.nio.channels.SocketChannel[connected local=/192.168.39.22:47700 > > remote= > > l-namenode2.dba.cn8.qunar.com/192.168.39.22:60000] > > at > > > > > org.apache.hadoop.hbase.ipc.RpcClient.callBlockingMethod(RpcClient.java:1667) > > at > > > > > org.apache.hadoop.hbase.ipc.RpcClient$BlockingRpcChannelImplementation.callBlockingMethod(RpcClient.java:1708) > > at > > > > > org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:40216) > > at > > > > > org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(HConnectionManager.java:1484) > > at > > > > > org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(HConnectionManager.java:2110) > > at > > > > > org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.getKeepAliveMasterService(HConnectionManager.java:1836) > > at > > > > > org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.listTables(HConnectionManager.java:2531) > > at > > org.apache.hadoop.hbase.client.HBaseAdmin.listTables(HBaseAdmin.java:298) > > at > > > > > org.apache.hadoop.hbase.tmpl.master.MasterStatusTmplImpl.__jamon_innerUnit__userTables(MasterStatusTmplImpl.java:530) > > at > > > > > org.apache.hadoop.hbase.tmpl.master.MasterStatusTmplImpl.renderNoFlush(MasterStatusTmplImpl.java:255) > > at > > > > > org.apache.hadoop.hbase.tmpl.master.MasterStatusTmpl.renderNoFlush(MasterStatusTmpl.java:382) > > at > > > > > org.apache.hadoop.hbase.tmpl.master.MasterStatusTmpl.render(MasterStatusTmpl.java:372) > > at > > > > > org.apache.hadoop.hbase.master.MasterStatusServlet.doGet(MasterStatusServlet.java:95) > > at javax.servlet.http.HttpServlet.service(HttpServlet.java:734) > > at javax.servlet.http.HttpServlet.service(HttpServlet.java:847) > > at org.mortbay.jetty.servlet.ServletHolder.handle(ServletHolder.java:511) > > at > > > > > org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1221) > > at > > > > > org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter.doFilter(StaticUserWebFilter.java:109) > > at > > > > > org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1212) > > at > > > > > org.apache.hadoop.http.HttpServer$QuotingInputFilter.doFilter(HttpServer.java:1081) > > at > > > > > org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1212) > > at org.apache.hadoop.http.NoCacheFilter.doFilter(NoCacheFilter.java:45) > > at > > > > > org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1212) > > at > org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:399) > > at > > > org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216) > > at > org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:182) > > at > org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:766) > > at org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:450) > > at > > > > > org.mortbay.jetty.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:230) > > at > org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:152) > > at org.mortbay.jetty.Server.handle(Server.java:326) > > at > org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:542) > > at > > > > > org.mortbay.jetty.HttpConnection$RequestHandler.headerComplete(HttpConnection.java:928) > > at org.mortbay.jetty.HttpParser.parseNext(HttpParser.java:549) > > at org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:212) > > at org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:404) > > at > > > > > org.mortbay.io.nio.SelectChannelEndPoint.run(SelectChannelEndPoint.java:410) > > at > > > > > org.mortbay.thread.QueuedThreadPool$PoolThread.run(QueuedThreadPool.java:582) > > Caused by: java.net.SocketTimeoutException: Call to > > l-namenode2.dba.cn8.qunar.com/192.168.39.22:60000 failed because > > java.net.SocketTimeoutException: 60000 millis timeout while waiting for > > channel to be ready for read. ch : > > java.nio.channels.SocketChannel[connected local=/192.168.39.22:47700 > > remote= > > l-namenode2.dba.cn8.qunar.com/192.168.39.22:60000] > > at > org.apache.hadoop.hbase.ipc.RpcClient.wrapException(RpcClient.java:1475) > > at org.apache.hadoop.hbase.ipc.RpcClient.call(RpcClient.java:1450) > > at > > > > > org.apache.hadoop.hbase.ipc.RpcClient.callBlockingMethod(RpcClient.java:1650) > > ... 37 more > > Caused by: java.net.SocketTimeoutException: 60000 millis timeout while > > waiting for channel to be ready for read. ch : > > java.nio.channels.SocketChannel[connected local=/192.168.39.22:47700 > > remote= > > l-namenode2.dba.cn8.qunar.com/192.168.39.22:60000] > > at > > > > > org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:164) > > at > org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161) > > at > org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:131) > > at java.io.FilterInputStream.read(FilterInputStream.java:133) > > at java.io.FilterInputStream.read(FilterInputStream.java:133) > > at > > > > > org.apache.hadoop.hbase.ipc.RpcClient$Connection$PingInputStream.read(RpcClient.java:553) > > at java.io.BufferedInputStream.fill(BufferedInputStream.java:235) > > at java.io.BufferedInputStream.read(BufferedInputStream.java:254) > > at java.io.DataInputStream.readInt(DataInputStream.java:387) > > at > > > > > org.apache.hadoop.hbase.ipc.RpcClient$Connection.readResponse(RpcClient.java:1057) > > at > org.apache.hadoop.hbase.ipc.RpcClient$Connection.run(RpcClient.java:719) > > Fri May 15 12:15:58 CST 2015 Terminating master > > {/logs} > > So what the exception means? Why? and how to solve the problem? > > >
