Yes, you should be able to make that work. -- Andrei Savu
On Tue, May 21, 2013 at 11:04 PM, Sebastien Goasguen <run...@gmail.com>wrote: > > On May 21, 2013, at 4:00 PM, Andrei Savu <savu.and...@gmail.com> wrote: > > You need sane dns settings (forward and reverse for each machine to make > this work). > > > Can I try to hack configure_hostname.sh in: > > services/cdh/target/classes/functions > > Adding some entry in /etc/hosts > > Will that be enough ? > > > -- Andrei Savu > > On Tue, May 21, 2013 at 10:57 PM, Sebastien Goasguen <run...@gmail.com>wrote: > >> >> On May 21, 2013, at 3:48 PM, Andrew Bayer <andrew.ba...@gmail.com> wrote: >> >> Yeah, DNS is a giant pain. If at all possible, you need to get the >> hostnames resolvable from wherever you're spinning the instances up, as >> well as on the instances themselves. The DNS that CloudStack's DHCP assigns >> should do the trick for that. >> >> >> argh… >> >> These instances have public IPs but not DNS entries. >> >> @andrei the hadoop-3d5 and other names are setup as the name of the >> instances. They are used for local 'hostname'. so no not resolvable. >> >> >> >> >> A. >> >> On Tue, May 21, 2013 at 12:46 PM, Sebastien Goasguen <run...@gmail.com>wrote: >> >>> Hi, >>> >>> I installed whirr 0.8.1, I am using it against a CloudStack endpoint. >>> Instances get launched and I am trying to setup cdh. >>> >>> I believe I am running into a DNS issue as I am running into lots of >>> issues of this type: >>> >>> 13/05/21 21:21:28 WARN net.DNS: Unable to determine local hostname >>> -falling back to "localhost" >>> java.net.UnknownHostException: hadoop-3d5: hadoop-3d5 >>> >>> If I log in to the name node and try to use hadoop I get things like: >>> >>> $ hadoop fs -mkdir /toto >>> -mkdir: java.net.UnknownHostException: hadoop-3d5 >>> >>> my hadoop-site.xml looks like: >>> >>> <?xml version="1.0"?> >>> <?xml-stylesheet type="text/xsl" href="configuration.xsl"?> >>> <configuration> >>> <property> >>> <name>dfs.client.use.legacy.blockreader</name> >>> <value>true</value> >>> </property> >>> <property> >>> <name>fs.default.name</name> >>> <value>hdfs://hadoop-3d5:8020/</value> >>> </property> >>> <property> >>> <name>mapred.job.tracker</name> >>> <value>hadoop-3d5:8021</value> >>> </property> >>> <property> >>> <name>hadoop.job.ugi</name> >>> <value>root,root</value> >>> </property> >>> <property> >>> <name>hadoop.rpc.socket.factory.class.default</name> >>> <value>org.apache.hadoop.net.SocksSocketFactory</value> >>> </property> >>> <property> >>> <name>hadoop.socks.server</name> >>> <value>localhost:6666</value> >>> </property> >>> </configuration> >>> >>> my ~/.whirr/hadoop/instances file has all the right IP addresses, but I >>> don't think the security group rules got created. >>> >>> Any thoughts ? >>> >>> thanks, >>> >>> -sebastien >>> >>> >> >> > >