Sounds like https://issues.apache.org/jira/browse/HBASE-3445 which has been committed on 0.90 branch and will be available in 0.90.1. St.Ack
On Tue, Feb 1, 2011 at 7:16 AM, grailcattt <[email protected]> wrote: > > I have hadoop/hbase running on a notebook as my dev env. I have everything > set up to use localhost which is defined as 127.0.0.1 in my /etc/hosts (and > removed other entries for localhost). > > The system works great all day, but when I go home and try to start the > system, it doesn't work with a different IP address. > > First thing I notice is from my namenode log is this cryptic INFO line: > <code> > 2011-02-01 07:56:09,696 INFO org.apache.hadoop.ipc.Server: IPC Server > handler 6 on 8020, call delete(/usr/share/hadoop/mapred/system, true) from > 127.0.0.1:49216: error: > org.apache.hadoop.hdfs.server.namenode.SafeModeException: Cannot delete > /usr/share/hadoop/mapred/system. Name node is in safe mode. > The ratio of reported blocks 0.0000 has not reached the threshold 0.9990. > Safe mode will be turned off automatically. > org.apache.hadoop.hdfs.server.namenode.SafeModeException: Cannot delete > /usr/share/hadoop/mapred/system. Name node is in safe mode. > The ratio of reported blocks 0.0000 has not reached the threshold 0.9990. > Safe mode will be turned off automatically. > at > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInternal(FSNamesystem.java:1700) > </code> > > Hadoop eventually exits safe mode, but this does not happen during startup > when the system is working. > > Next thing I notice upon starting hbase, in my namenode log: > <code> > 2011-02-01 08:04:34,313 INFO > org.apache.hadoop.hdfs.server.namenode.FSNamesystem.audit: > ugi=pansonm,staff,hadoop,com.apple.access_screensharing,_developer,_lpoperator,_lpadmin,_appserveradm,admin,_appserverusr,localaccounts,everyone,hadoop1,com.apple.sharepoint.group.1,com.apple.sharepoint.group.2 > ip=/127.0.0.1 cmd=mkdirs > src=/hbase/.logs/192.168.1.12,49320,1296572670348 > dst=null perm=pansonm:supergroup:rwxr-xr-x > </code> > > NOTICE: The reference to my LAN IP address 192.168.1.12. This isn't an > error, but it is curious that hadoop is not using localhost/127.0.0.1 > somewhere. > > The main problem appears in my hbase log: > > <code> > 2011-02-01 08:04:38,900 DEBUG > org.apache.hadoop.hbase.master.handler.OpenedRegionHandler: Opened region > -ROOT-,,0.70236052 on 192.168.1.12,49320,1296572670348 > 2011-02-01 08:04:58,931 INFO org.apache.hadoop.ipc.HbaseRPC: Problem > connecting to server: 192.168.1.2/192.168.1.2:51038 > 2011-02-01 08:04:59,934 FATAL org.apache.hadoop.hbase.master.HMaster: > Unhandled exception. Starting shutdown. > java.net.SocketException: Host is down > at sun.nio.ch.Net.connect(Native Method) > at sun.nio.ch.SocketChannelImpl.connect(SocketChannelImpl.java:507) > at > org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:192) > </code> > > The reference to 192.168.1.2 certainly won't work since this is my old > address. And it appears that this hard-coded address is used as a locator > for data. > > The only solution I have now is deleting all the data and reformatting HDFS > -- which I'm now doing twice per day. > > Thanks much for your help. > -- > View this message in context: > http://old.nabble.com/HBase-Zookeeper----System-Fails-when-IP-Address-Changes-tp30816966p30816966.html > Sent from the HBase User mailing list archive at Nabble.com. > >
