Thanks again. This changes the behavior, but it does not yet fix my problem. The hbase.rootdir property forces the hbase master to stay alive for a little while, so I had a moment of short-lived euphoria when Hmaster appeared in the jps list, but this only lasts while it tries to connect to localhost:9000 (which is not open), and it still doesn't open port 60000 and it still thinks it is named my-static-ip.com (i.e., same error message as before). The removal of localhost.localdomain from /etc/hosts made no difference one way or the other. I still am looking for a way to try to have hbase bind to localhost:6000 instead of my-static-ip.com:6000. I will also try to see why localhost:9000 is not open (though that appears later in the log file, so I don't think it is causing the failure to open 60000). Thanks for the help so far, I will post again with further info.
Michael On Thu, Sep 16, 2010 at 12:53 PM, N.N. Gesli <[email protected]> wrote: > I have this in hbase-site.xml: > > <property> > <name>hbase.rootdir</name> > <value>hdfs://localhost:9000/hbase</value> > <description>The directory shared by region servers. > Should be fully-qualified to include the filesystem to use. > E.g: hdfs://NAMENODE_SERVER:PORT/HBASE_ROOTDIR > </description> > </property> > > <property> > <name>hbase.cluster.distributed</name> > <value>true</value> > <description>For psuedo-distributed, you want to set this to true. > false means that HBase tries to put Master + RegionServers in one > process. > Pseudo-distributed = seperate processes/pids</description> > </property> > > <property> > <name>hbase.regionserver.hlog.replication</name> > <value>1</value> > <description>For HBase to offer good data durability, we roll logs if > filesystem replication falls below a certain amount. In > psuedo-distributed > mode, you normally only have the local filesystem or 1 HDFS DataNode, so > you > don't want to roll logs constantly.</description> > </property> > > <property> > <name>hbase.tmp.dir</name> > <value>/tmp/hbase-testing</value> > <description>Temporary directory on the local filesystem.</description> > </property> > > I also hase Hadoop conf directory in HBASE_CLASSPATH (hbase-env.sh). > > I just tried etc/hosts with "127.0.0.1 localhost.localdomain > localhost" line. I got the same error I was getting before. I switched it > back to "127.0.0.1 localhost" and it worked. In between those > changes, > I stopped hbase, hadoop and killed still running region server. I hope that > helps. > > N.Gesli > > > > > On Thu, Sep 16, 2010 at 7:04 AM, Michael Scott <[email protected]> > wrote: > > > This sounds promising, I have one quick question about your steps: where > > in > > the Hbase config *site*.xml did you make the change back to localhost? > My > > hbase master is using the public IP address (97.86.88.18), and I don't > > think > > I've told it to. I want to convince hbase to get rid of the line in the > > log > > file that says something like: > > > > 2010-09-16 09:59:21,727 INFO org.apache.hadoop.hbase.master.HMaster: My > > address is 97-86-88-18.static.aldl.mi.charter.com:60000 > > > > (Note that my /etc/hosts has only the one line > > 127.0.0.1 localhost.localdomain localhost > > since I'm not running ipv6, but somehow hbase knows that the interface is > a > > comcast static address. I can use /etc/hosts to change that to the > > registered domain name for 97-86-88-18, but this doesn't help.) > > > > To reply to Ryan's question, my ifconfig gives: > > > > eth0 Link encap:Ethernet HWaddr 00:24:E8:01:DA:B8 > > inet addr:10.0.0.2 Bcast:10.0.0.255 Mask:255.255.255.0 > > UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 > > RX packets:319475 errors:0 dropped:0 overruns:0 frame:0 > > TX packets:290698 errors:0 dropped:0 overruns:0 carrier:0 > > collisions:0 txqueuelen:1000 > > RX bytes:108186958 (103.1 MiB) TX bytes:187845633 (179.1 MiB) > > Interrupt:28 Base address:0xa000 > > > > lo Link encap:Local Loopback > > inet addr:127.0.0.1 Mask:255.0.0.0 > > UP LOOPBACK RUNNING MTU:16436 Metric:1 > > RX packets:370795 errors:0 dropped:0 overruns:0 frame:0 > > TX packets:370795 errors:0 dropped:0 overruns:0 carrier:0 > > collisions:0 txqueuelen:0 > > RX bytes:108117402 (103.1 MiB) TX bytes:108117402 (103.1 MiB) > > > > Thanks a bunch! > > > > Michael > > > > On Thu, Sep 16, 2010 at 1:12 AM, N.N. Gesli <[email protected]> wrote: > > > > > Hi Michael, > > > > > > I was having a similar problem and following this thread for any > > > suggestions. I tried everything suggested and more. > > > > > > I was trying to run Hadoop/Hbase pseudo distributed version on my Mac. > I > > > initially started with Hadoop 21.0 and Hbase 0.89 versions. I had > exactly > > > the same error that you were getting. Then switched to Hadoop 20.2 and > > > Hbase > > > 20.6 - still HMaster was not starting. Then finally it worked. Below > are > > my > > > steps to success :) > > > > > > * stopped hbase > > > * stopped hadoop > > > * run jps; RegionServer was still running; killed it manually > > > * in tmp directory (where hadoop namenode and *.pid files are stored) I > > > removed everything related to hadoop and hbase, including the > > directories. > > > (I had no data in Hadoop, so I could do this) > > > * changed the ports back to default 600** > > > * changed back Hadoop and Hbase configurations to "localhost" in > > *site*.xml > > > and regionservers. (Only I will be using this - no remote connection) > > > * changed back my /etc/hosts to the original version. It looks like > this: > > > 127.0.0.1 localhost > > > ::1 localhost > > > fe80::1%lo0 localhost > > > * reformatted the Hadoop namenode > > > * started Hadoop > > > * started HBase and it worked :) > > > > > > Let me know if you want to know any specific configuration. > > > > > > N.Gesli > > > > > > On Wed, Sep 15, 2010 at 10:41 PM, Ryan Rawson <[email protected]> > > wrote: > > > > > > > What is your ifconfig output looking like? > > > > > > > > > > > > > > > > On Wed, Sep 15, 2010 at 10:07 PM, Michael Scott < > [email protected]> > > > > wrote: > > > > > Thanks for the continued advice. I am still confused by the > > different > > > > > behaviors of hadoop and hbase. As I said before, I can't get hbase > to > > > > work > > > > > on any of the ports that hadoop works on, so I guess hadoop and > hbase > > > are > > > > > using different interfaces. Why is this, and can't I ask hbase to > > use > > > > the > > > > > interface that hadoop uses? What interfaces are hadoop and hbase > > > using? > > > > > > > > > > Also (and maybe this is the wrong forum for this question), how can > I > > > get > > > > my > > > > > OS to allow me to open 60000 using the IP address? I have > > temporarily > > > > > disabled selinux and iptables, as I thought that this would simply > > > allow > > > > all > > > > > port connections. Still, this works just fine: > > > > > bash-4.0$ nc -l 60000 > /tmp/nc.out > > > > > > > > > > but this does not: > > > > > bash-4.0$ nc -l 97.86.88.18 60000 > /tmp/nc.out > > > > > (returns "nc: Cannot assign requested address"; I get the same > error > > > for > > > > the > > > > > hostname instead of the IP address, and for 10.0.0.1, but 10.0.0.0 > is > > > > > allowed) > > > > > > > > > > I am trying to get hbase running for a socorro server, which will > > > running > > > > > locally. I don't know if that matters. > > > > > > > > > > Thanks, > > > > > Michael > > > > > > > > > > On Wed, Sep 15, 2010 at 6:04 PM, Ryan Rawson <[email protected]> > > > wrote: > > > > > > > > > >> Hey, > > > > >> > > > > >> If you bind to localhost you wont actually be reachable by anyone! > > > > >> > > > > >> The question is why is your OS disallowing binds to a specific > > > > >> interface/port combo? > > > > >> > > > > >> HBase does not really run in a blended/multihomed environment... > > > > >> meaning if you have multiple interfaces, you have to choose one > that > > > > >> we work over. This is because we need to know a singular > canonical > > > > >> IP/name for any given server because we put that info up inside > > > > >> ZooKeeper and META tables. So it's not just an artificial > > constraint, > > > > >> but exists for cluster management needs. > > > > >> > > > > >> Having said that, we do work on multihomed machines, eg: ec2, you > > > > >> might bind hbase to the internal interface taking advantage of the > > > > >> unmetered/faster network. Also better for security as well. > > > > >> > > > > >> Let us know if you need more background on how we use the network > > and > > > > why. > > > > >> -ryan > > > > >> > > > > >> On Wed, Sep 15, 2010 at 10:18 AM, Michael Scott < > > [email protected] > > > > > > > > >> wrote: > > > > >> > Hi again, > > > > >> > > > > > >> > I think the hbase server master is not starting because it is > > > > attempting > > > > >> to > > > > >> > open port 60000 on its public IP address, rather than using > > > localhost. > > > > I > > > > >> > cannot seem to figure out how to force it (well, configure it) > to > > > > attempt > > > > >> to > > > > >> > bind to localhost:60000 instead. As far as I can see, this is > > set > > > in > > > > >> the > > > > >> > file: > > > > >> > > > > > >> > org/apache/hadoop/hbase/master/HMaster.java > > > > >> > > > > > >> > I don't know much about java, so I'd prefer not to edit the > source > > > if > > > > >> there > > > > >> > is an option, but I will if necessary. Can someone please point > > me > > > to > > > > >> the > > > > >> > way to change this setting? Any help would be greatly > > appreciated. > > > > >> > > > > > >> > Thanks, > > > > >> > Michael > > > > >> > > > > > >> > On Wed, Sep 15, 2010 at 12:42 AM, Michael Scott < > > > [email protected] > > > > >> >wrote: > > > > >> > > > > > >> >> Hi again, > > > > >> >> > > > > >> >> IPV6 was enabled. I shut it off, rebooted to be sure, verified > > it > > > > was > > > > >> >> still off, and encountered the same problem once again. > > > > >> >> > > > > >> >> I also tried to open port 60000 by hand with a small php file. > I > > > can > > > > do > > > > >> >> this (as any user) for localhost. I can NOT do this (not even > as > > > > root) > > > > >> for > > > > >> >> the IP address which matches the fully qualified domain name, > > which > > > > is > > > > >> what > > > > >> >> hbase is trying to use. Is there some way for me to configure > > > hbase > > > > to > > > > >> use > > > > >> >> localhost instead of the fully qualified domain name for the > > > master? > > > > I > > > > >> >> would have thought this was done by default, or that there > would > > be > > > > an > > > > >> >> obvious line in some conf file, but I can't find it. > > > > >> >> > > > > >> >> Thanks again, > > > > >> >> > > > > >> >> Michael > > > > >> >> > > > > >> >> On Tue, Sep 14, 2010 at 12:23 PM, Todd Lipcon < > [email protected] > > > > > > > >> wrote: > > > > >> >> > > > > >> >>> Hi Michael, > > > > >> >>> > > > > >> >>> It might be related to IPV6. Do you have IPV6 enabled on this > > > > machine? > > > > >> >>> > > > > >> >>> Check out this hadoop JIRA that might be related for some > tips: > > > > >> >>> https://issues.apache.org/jira/browse/HADOOP-6056 > > > > >> >>> > > > > >> >>> <https://issues.apache.org/jira/browse/HADOOP-6056>-Todd > > > > >> >>> > > > > >> >>> On Tue, Sep 14, 2010 at 10:17 AM, Michael Scott < > > > > [email protected] > > > > >> >>> >wrote: > > > > >> >>> > > > > >> >>> > That's correct. I tried a number of different ports to see > if > > > > there > > > > >> was > > > > >> >>> > something weird, and then I shut down the hadoop server and > > > tried > > > > to > > > > >> >>> > connect > > > > >> >>> > to 50010 (which of course should have been free at that > point) > > > but > > > > >> got > > > > >> >>> the > > > > >> >>> > same "cannot assign to requested address" error. If I start > > > > hadoop, > > > > >> >>> > netstat > > > > >> >>> > shows a process listening on 50010. > > > > >> >>> > > > > > >> >>> > I am going to try this on a different OS, I am wondering if > > FC11 > > > > is > > > > >> my > > > > >> >>> > problem. > > > > >> >>> > > > > > >> >>> > Michael > > > > >> >>> > > > > > >> >>> > On Tue, Sep 14, 2010 at 11:41 AM, Stack <[email protected]> > > > wrote: > > > > >> >>> > > > > > >> >>> > > On Tue, Sep 14, 2010 at 9:33 AM, Michael Scott < > > > > >> [email protected]> > > > > >> >>> > > wrote: > > > > >> >>> > > > I don't see why hadoop binds > > > > >> >>> > > > to a port but hbase does not (I even tried starting > hbase > > > with > > > > >> >>> hadoop > > > > >> >>> > off > > > > >> >>> > > > and binding to 50010, which hadoop uses). > > > > >> >>> > > > > > > > >> >>> > > > > > > >> >>> > > Using 50010 worked for hadoop but not for hbase? (Odd. > We > > > > hadoop > > > > >> >>> > > their mechanism essentially). > > > > >> >>> > > > > > > >> >>> > > St.Ack > > > > >> >>> > > > > > > >> >>> > > > > > >> >>> > > > > >> >>> > > > > >> >>> > > > > >> >>> -- > > > > >> >>> Todd Lipcon > > > > >> >>> Software Engineer, Cloudera > > > > >> >>> > > > > >> >> > > > > >> >> > > > > >> > > > > > >> > > > > > > > > > > > > > > >
