Thanks Philippe.

Yeah, sorry, I should I have mentioned that I tried using the hostname
of my machine first, so I had the following hadoop-site.xml settings.

<property>
 <name>fs.default.name</name>
 <value>tiger.stanford.edu:9000</value>
</property>

<!-- map/reduce properties -->

<property>
 <name>mapred.job.tracker</name>
 <value>tiger.stanford.edu:9001</value>
</property>

But that still didn't work:

tiger$ bin/hadoop jar hadoop-0.12.2-examples.jar wordcount input_dir output_dir

07/03/26 01:57:25 INFO ipc.Client: Retrying connect to server:
tiger.stanford.edu/
xx.yy.zz.aa:9000. Already tried 1 time(s).
...
xx.yy.zz.aa:9000. Already tried 10 time(s).
java.lang.RuntimeException: java.net.ConnectException: Connection refused

Separately Arun - I did have passphrase-less ssh enabled on this machine.

i.e., I executed:

ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa
cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys

On 3/26/07, Philippe Gassmann <[EMAIL PROTECTED]> wrote:
Hi,

Tiger Uppercut a écrit :
> <snip/>
>
> <property>
>  <name>fs.default.name</name>
>  <value>localhost:9000</value>
> </property>
>
> <!-- map/reduce properties -->
>
> <property>
>  <name>mapred.job.tracker</name>
>  <value>localhost:9001</value>
> </property>
>
For the fs.default.name and the mapred.job.tracker try to use the
hostname of your machine instead of localhost. When using
localhost:XXXX, hadoop servers are listen to the loopback interface. But
mapreduce jobs (I do not know exactly where) are seeing that the
connections to tasktrackers are issued using the 127.0.0.1 and are
trying to reverse dns the adress. Your system will not return localhost
but the real name of your machine. In most linux system, that name is
binded to an ethernet interface so jobs will try to connect to that
interface instead of the loopback one.



> <property>
>  <name>dfs.name.dir</name>
>  <value>/some_dir/hadoop/hadoop_data</value>
> </property>
>
> </configuration>


Reply via email to