[ https://issues.apache.org/jira/browse/HBASE-1973?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
stack resolved HBASE-1973. -------------------------- Resolution: Fixed Apply HDFS-630 to fix this issue. Resolving. > Second RS goes down because "Bad connect ack with firstBadLink as..." > --------------------------------------------------------------------- > > Key: HBASE-1973 > URL: https://issues.apache.org/jira/browse/HBASE-1973 > Project: Hadoop HBase > Issue Type: Bug > Reporter: stack > > Placeholder for issue where a second RS goes down when I kill another RS and > its DN. I asked over on hdfs-user why we're stuck on 140. Why can't it move > on to another DN creating a block. > {code} > 2009-11-12 05:09:38,624 [regionserver/208.76.44.141:60020.logRoller] INFO > org.apache.hadoop.hbase.regionserver.wal.HLog: Roll > /hbase/.logs/aa0-000-14.u.powerset.com,60020,1258002404935/hlog.dat.1258002405346, > entries=52747, calcsize=63757973, filesize=58588977. New hlog > /hbase/.logs/aa0-000-14.u.powerset.com,60020,12 > 58002404935/hlog.dat.1258002578617 > 2009-11-12 05:09:38,628 [regionserver/208.76.44.141:60020.logRoller] DEBUG > org.apache.hadoop.hbase.regionserver.wal.HLog: Found 0 hlogs to remove out > of total 1; oldest outstanding seqnum is 1 from region .META.,,1 > 2009-11-12 05:09:38,628 [IPC Server handler 16 on 60020] INFO > org.apache.hadoop.hbase.regionserver.wal.HLog: edit=0, > write=TestTable/TestTable,,1258002466308/612228 > 2009-11-12 05:09:38,630 [regionserver/208.76.44.141:60020.logSyncer] INFO > org.apache.hadoop.hbase.regionserver.wal.HLog: sync > 2009-11-12 05:09:38,646 [Thread-61] INFO org.apache.hadoop.hdfs.DFSClient: > Exception in createBlockOutputStream java.io.IOException: Bad connect ack > with firstBadLink as 208.76.44.140:51010 > 2009-11-12 05:09:38,646 [Thread-61] INFO org.apache.hadoop.hdfs.DFSClient: > Abandoning block blk_1786802275426334496_1076 > 2009-11-12 05:09:44,662 [Thread-61] INFO org.apache.hadoop.hdfs.DFSClient: > Exception in createBlockOutputStream java.io.IOException: Bad connect ack > with firstBadLink as 208.76.44.140:51010 > 2009-11-12 05:09:44,662 [Thread-61] INFO org.apache.hadoop.hdfs.DFSClient: > Abandoning block blk_5838454320666652488_1076 > 2009-11-12 05:09:46,226 [pool-1-thread-1] DEBUG > org.apache.hadoop.hbase.io.hfile.LruBlockCache: Cache Stats: Sizes: > Total=5.5656967MB (5836056), Free=672.77185MB (705452392), Max=678.3375MB > (711288448), Counts: Blocks=0, Access=0, Hit=0, Miss=0, Evictions=0, > Evicted=0, Ratios: Hit Ratio=NaN%, Miss Ratio=NaN%, Evicted > /Run=NaN > 2009-11-12 05:09:50,683 [Thread-61] INFO org.apache.hadoop.hdfs.DFSClient: > Exception in createBlockOutputStream java.io.IOException: Bad connect ack > with firstBadLink as 208.76.44.140:51010 > 2009-11-12 05:09:50,684 [Thread-61] INFO org.apache.hadoop.hdfs.DFSClient: > Abandoning block blk_8615264034947520229_1076 > 2009-11-12 05:09:56,691 [Thread-61] INFO org.apache.hadoop.hdfs.DFSClient: > Exception in createBlockOutputStream java.io.IOException: Bad connect ack > with firstBadLink as 208.76.44.140:51010 > 2009-11-12 05:09:56,692 [Thread-61] INFO org.apache.hadoop.hdfs.DFSClient: > Abandoning block blk_-5632723632070749366_1077 > 2009-11-12 05:10:02,696 [Thread-61] WARN org.apache.hadoop.hdfs.DFSClient: > DataStreamer Exception: java.io.IOException: Unable to create new block. > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSClient.java:3100) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2681) > 2009-11-12 05:10:02,697 [Thread-61] WARN org.apache.hadoop.hdfs.DFSClient: > Could not get block locations. Source file > "/hbase/.logs/aa0-000-14.u.powerset.com,60020,1258002404935/hlog.dat.1258002578617" > - Aborting... > 2009-11-12 05:10:02,699 [regionserver/208.76.44.141:60020.logSyncer] FATAL > org.apache.hadoop.hbase.regionserver.wal.HLog: Could not append. Requesting > close of hlog > java.io.IOException: Bad connect ack with firstBadLink as 208.76.44.140:51010 > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.createBlockOutputStream(DFSClient.java:3160) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSClient.java:3080) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2681) > 2009-11-12 05:10:02,701 [regionserver/208.76.44.141:60020.logSyncer] ERROR > org.apache.hadoop.hbase.regionserver.wal.HLog: Error while syncing, > requesting close of hlog > java.io.IOException: Bad connect ack with firstBadLink as 208.76.44.140:51010 > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.createBlockOutputStream(DFSClient.java:3160) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSClient.java:3080) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream$DataStreamer.run(DFSClient.java:2681) > 2009-11-12 05:10:02,703 [IPC Server handler 20 on 60020] FATAL > org.apache.hadoop.hbase.regionserver.wal.HLog: Could not append. Requesting > close of hlog > java.io.IOException: IOException flush:java.io.IOException: Bad connect ack > with firstBadLink as 208.76.44.140:51010 > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.hflush(DFSClient.java:3527) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.sync(DFSClient.java:3473) > at > org.apache.hadoop.fs.FSDataOutputStream.sync(FSDataOutputStream.java:97) > at org.apache.hadoop.hbase.regionserver.wal.HLog.hflush(HLog.java:829) > at > org.apache.hadoop.hbase.regionserver.wal.HLog$LogSyncer.run(HLog.java:751) > 2009-11-12 05:10:02,703 [regionserver/208.76.44.141:60020.logSyncer] INFO > org.apache.hadoop.hbase.regionserver.wal.HLog: > regionserver/208.76.44.141:60020.logSyncer exiting > 2009-11-12 05:10:02,706 [IPC Server handler 10 on 60020] FATAL > org.apache.hadoop.hbase.regionserver.wal.HLog: Could not append. Requesting > close of hlog > java.io.IOException: IOException flush:java.io.IOException: Bad connect ack > with firstBadLink as 208.76.44.140:51010 > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.hflush(DFSClient.java:3527) > at > org.apache.hadoop.hdfs.DFSClient$DFSOutputStream.sync(DFSClient.java:3473) > at > org.apache.hadoop.fs.FSDataOutputStream.sync(FSDataOutputStream.java:97) > at org.apache.hadoop.hbase.regionserver.wal.HLog.hflush(HLog.java:829) > at > org.apache.hadoop.hbase.regionserver.wal.HLog$LogSyncer.run(HLog.java:751) > {code} -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.