One thing i notice is that /etc/hosts is different on each host: amlan-laptop 
is bound to localhost on the master and its bound to a different ip on the 
slave. Make the files on both macnes the same.

Sent from my iPad

On Feb 21, 2011, at 2:06, Amlan Mandal <am...@fourint.com> wrote:

> Thanks MIS.
> 
> Can somebody please tell me what is wrong with this.
> 
> 
> cat /etc/hosts (on master)
> 
> 127.0.0.1   localhost   amlan-laptop
> 192.168.1.11    dhan
> 
> 
> cat /etc/hosts (on slave)
> 127.0.0.1       localhost       dhan
> 192.168.1.22    amlan-laptop
> 
> cat conf/masters (on master)
> amlan-laptop
> 
> cat conf/slaves
> amlan-laptop
> dhan
> 
> 
> cat conf/core-site.xml on BOTH machines
> <name>fs.default.name</name>
>                 <value>hdfs://amlan-laptop:54310</value>
> 
> cat conf/mapred-site.xml on BOTH machines
> <name>mapred.job.tracker</name>
>         <value>amlan-laptop:54311</value>
> 
> 
> hostname (on master)
> amlan-laptop
> 
> hostname (on slave)
> dhan
> 
> 
> passwordless SSH from master TO master works fine (ssh amlan-laptop)
> passwordless SSH from master TO slave works fine (ssh dhan)
> 
> 
> I see on slave datanode log when start-dfs.sh
> 
> 
> 2011-02-21 15:25:31,304 INFO org.apache.hadoop.ipc.RPC: Server at 
> amlan-laptop/192.168.1.22:54310 not available yet, Zzzzz...
> 2011-02-21 15:25:33,312 INFO org.apache.hadoop.ipc.Client: Retrying connect 
> to server: amlan-laptop/192.168.1.22:54310. Already tried 0 time(s).
> 
> 
> Why on the earth amlan-laptop/192.168.1.22:54310  ?????
> 
> 
> should NOT it be amlan-laptop:54310 ???
> 
> why does it concatenate hostname/ip ???
> 
> Can somebody PLEASE help me out.
> 
> 
> On Mon, Feb 21, 2011 at 2:12 PM, MIS <misapa...@gmail.com> wrote:
> Please have the host-name and ip address mapping in the /etc/hosts file on 
> both the nodes that are running hadoop cluster. 
> 
> One more thing  : I hope secondary namenode is also running along namenode 
> but you may have forgot to mention it.
> 
> Thanks,
> MIS
> 
> 
> On Mon, Feb 21, 2011 at 12:47 PM, Amlan Mandal <am...@fourint.com> wrote:
> Thanks Mafish.
> Can you please point me which config need to be set correctly?
> 
> Amlan
> 
> 
> On Mon, Feb 21, 2011 at 12:45 PM, Mafish Liu <maf...@gmail.com> wrote:
> It seem you did not config your HDFS properly.
> 
> "Caused by: java.lang.IllegalArgumentException: Wrong FS:
> hdfs://192.168.1.22:54310/tmp/hive-hadoop/hive_2011-02-21_12-09-42_678_6107747797061030113,
> expected: hdfs://amlan-laptop.local:54310 "
> 
> 
> 
> 2011/2/21 Amlan Mandal <am...@fourint.com>:
> > To give more context my multinode hadoop is working fine. fs.default.name,
> > mapred.job.tracker settings are correct.
> > I can submit job to my multinode hadoop and see output.  (One of the node
> > running namenode,datanode,job tracker , task tracker other running task
> > tracker and datanode)
> >
> > On Mon, Feb 21, 2011 at 12:24 PM, Amlan Mandal <am...@fourint.com> wrote:
> >>
> >> Earlier I had hive running on single node hadoop which was working fine.
> >> Now I made it 2 node hadoop cluster. When I run hive from cli I am getting
> >> following error
> >>
> >>
> >> java.lang.RuntimeException: Error while making MR scratch directory -
> >> check filesystem config (null)
> >>     at org.apache.hadoop.hive.ql.Context.getMRScratchDir(Context.java:216)
> >>     at org.apache.hadoop.hive.ql.Context.getMRTmpFileURI(Context.java:292)
> >>     at
> >> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.getMetaData(SemanticAnalyzer.java:825)
> >>     at
> >> org.apache.hadoop.hive.ql.parse.SemanticAnalyzer.analyzeInternal(SemanticAnalyzer.java:6093)
> >>     at
> >> org.apache.hadoop.hive.ql.parse.BaseSemanticAnalyzer.analyze(BaseSemanticAnalyzer.java:125)
> >>     at org.apache.hadoop.hive.ql.Driver.compile(Driver.java:304)
> >>     at org.apache.hadoop.hive.ql.Driver.run(Driver.java:379)
> >>     at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:138)
> >>     at
> >> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:197)
> >>     at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:302)
> >>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> >>     at
> >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> >>     at
> >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> >>     at java.lang.reflect.Method.invoke(Method.java:597)
> >>     at org.apache.hadoop.util.RunJar.main(RunJar.java:156)
> >> Caused by: java.lang.IllegalArgumentException: Wrong FS:
> >> hdfs://192.168.1.22:54310/tmp/hive-hadoop/hive_2011-02-21_12-09-42_678_6107747797061030113,
> >> expected: hdfs://amlan-laptop.local:54310
> >> ...
> >>
> >>
> >> I can guess I need to change some config variable for hive , can somebody
> >> please help me out?
> >>
> >>
> >> Amlan
> >
> >
> 
> 
> 

Reply via email to